From Digital Age to Nano Age. WorldWide.

Tag: Google Cloud Next 2024

Robotic Automations

Google's Gemini comes to databases | TechCrunch


Google wants Gemini, its family of generative AI models, to power your app’s databases — in a sense.

At its annual Cloud Next conference in Las Vegas, Google announced the public preview of Gemini in Databases, a collection of features underpinned by Gemini to — as the company pitched it — “simplify all aspects of the database journey.” In less jargony language, Gemini in Databases is a bundle of AI-powered, developer-focused tools for Google Cloud customers who are creating, monitoring and migrating app databases.

One piece of Gemini in Databases is Database Studio, an editor for structured query language (SQL), the language used to store and process data in relational databases. Built into the Google Cloud console, Database Studio can generate, summarize and fix certain errors with SQL code, Google says, in addition to offering general SQL coding suggestions through a chatbot-like interface.

Joining Database Studio under the Gemini in Databases brand umbrella is AI-assisted migrations via Google’s existing Database Migration Service. Google’s Gemini models can convert database code and deliver explanations of those changes along with recommendations, according to Google.

Elsewhere, in Google’s new Database Center — yet another Gemini in Databases component — users can interact with databases using natural language and can manage a fleet of databases with tools to assess their availability, security and privacy compliance. And should something go wrong, those users can ask a Gemini-powered bot to offer troubleshooting tips.

“Gemini in Databases enables customer to easily generate SQL; additionally, they can now manage, optimize and govern entire fleets of databases from a single pane of glass; and finally, accelerate database migrations with AI-assisted code conversions,” Andi Gutmans, GM of databases at Google Cloud, wrote in a blog post shared with TechCrunch. “Imagine being able to ask questions like ‘Which of my production databases in east Asia had missing backups in the last 24 hours?’ or ‘How many PostgreSQL resources have a version higher than 11?’ and getting instant insights about your entire database fleet.”

That assumes, of course, that the Gemini models don’t make mistakes from time to time — which is no guarantee.

Regardless, Google’s forging ahead, bringing Gemini to Looker, its business intelligence tool, as well.

Launching in private preview, Gemini in Looker lets users “chat with their business data,” as Google describes it in a blog post. Integrated with Workspace, Google’s suite of enterprise productivity tools, Gemini in Looker spans features such as conversational analytics; report, visualization and formula generation; and automated Google Slide presentation generation. 

I’m curious to see if Gemini in Looker’s report and presentation generation work reliably well. Generative AI models don’t exactly have a reputation for accuracy, after all, which could lead to embarrassing, or even mission-critical, mistakes. We’ll find out as Cloud Next continues into the week with any luck.

Gemini in Databases could be perceived as a response of sorts to top rival Microsoft’s recently launched Copilot in Azure SQL Database, which brought generative AI to Microsoft’s existing fully managed cloud database service. Microsoft is looking to stay a step ahead in the budding AI-driven database race and has also worked to build generative AI with Azure Data Studio, the company’s set of enterprise data management and development tools.


Software Development in Sri Lanka

Robotic Automations

Google open sources tools to support AI model development | TechCrunch


In a typical year, Cloud Next — one of Google’s two major annual developer conferences, the other being I/O — almost exclusively features managed and otherwise closed source, gated-behind-locked-down-APIs products and services. But this year, whether to foster developer goodwill or advance its ecosystem ambitions (or both), Google debuted a number of open source tools primarily aimed at supporting generative AI projects and infrastructure.

The first, MaxDiffusion, which Google actually quietly released in February, is a collection of reference implementations of various diffusion models — models like the image generator Stable Diffusion — that run on XLA devices. “XLA” stands for Accelerated Linear Algebra, an admittedly awkward acronym referring to a technique that optimizes and speeds up specific types of AI workloads, including fine-tuning and serving.

Google’s own tensor processing units (TPUs) are XLA devices, as are recent Nvidia GPUs.

Beyond MaxDiffusion, Google’s launching JetStream, a new engine to run generative AI models — specifically text-generating models (so not Stable Diffusion). Currently limited to supporting TPUs with GPU compatibility supposedly coming in the future, JetStream offers up to 3x higher “performance per dollar” for models like Google’s own Gemma 7B and Meta’s Llama 2, Google claims.

“As customers bring their AI workloads to production, there’s an increasing demand for a cost-efficient inference stack that delivers high performance,” Mark Lohmeyer, Google Cloud’s GM of compute and machine learning infrastructure, wrote in a blog post shared with TechCrunch. “JetStream helps with this need … and includes optimizations for popular open models such as Llama 2 and Gemma.”

Now, “3x” improvement is quite a claim to make, and it’s not exactly clear how Google arrived at that figure. Using which generation of TPU? Compared to which baseline engine? And how’s “performance” being defined here, anyway?

I’ve asked Google all these questions and will update this post if I hear back.

Second-to-last on the list of Google’s open source contributions are new additions to MaxText, Google’s collection of text-generating AI models targeting TPUs and Nvidia GPUs in the cloud. MaxText now includes Gemma 7B, OpenAI’s GPT-3 (the predecessor to GPT-4), Llama 2 and models from AI startup Mistral — all of which Google says can be customized and fine-tuned to developers’ needs.

We’ve heavily optimized [the models’] performance on TPUs and also partnered closely with Nvidia to optimize performance on large GPU clusters,” Lohmeyer said. “These improvements maximize GPU and TPU utilization, leading to higher energy efficiency and cost optimization.”

Finally, Google’s collaborated with Hugging Face, the AI startup, to create Optimum TPU, which provides tooling to bring certain AI workloads to TPUs. The goal is to reduce the barrier to entry for getting generative AI models onto TPU hardware, according to Google — in particular text-generating models.

But at present, Optimum TPU is a bit bare-bones. The only model it works with is Gemma 7B. And Optimum TPU doesn’t yet support training generative models on TPUs — only running them.

Google’s promising improvements down the line.


Software Development in Sri Lanka

Robotic Automations

Google announces Axion, its first custom Arm-based data center processor | TechCrunch


Google Cloud on Tuesday joined AWS and Azure in announcing its first custom-built Arm processor, dubbed Axion. Based on Arm’s Neoverse 2 designs, Google says its Axion instances offer 30% better performance than other Arm-based instances from competitors like AWS and Microsoft and up to 50% better performance and 60% better energy efficiency than comparable X86-based instances.

Google did not provide any documentation to back these claims up and, like us, you’d probably like to know more about these chips. We asked a lot of questions, but Google politely declined to provide any additional information. No availability dates, no pricing, no additional technical data. Those “benchmark” results? The company wouldn’t even say which X86 instance it was comparing Axion to.

“Technical documentation, including benchmarking and architecture details, will be available later this year,” Google spokesperson Amanda Lam said.

Image Credits: Frederic Lardinois/TechCrunch

Maybe the chips aren’t even ready yet? After all, it took Google a while to announce Arm-chips in the cloud, especially considering that Google has long built its in-house TPU AI chips and, more recently, custom Arm-based mobile chips for its Pixel phones. AWS launched its Graviton chips back in 2018.

To be fair, though, Microsoft only announced its Cobalt Arm chips late last year, too, and those chips aren’t yet available to customers, either. But Microsoft Azure has offered instances based on Ampere’s Arm servers since 2022.

In a press briefing ahead of Tuesday’s announcement, Google stressed that since Axion is built on an open foundation, Google Cloud customers will be able to bring their existing Arm workloads to Google Cloud without any modifications. That’s really no surprise. Anything else would’ve been a very dumb move on Google Cloud’s part.

Image Credits: Frederic Lardinois/TechCrunch

“We recently contributed to the SystemReady Virtual Environment, which is Arm’s hardware and firmware interoperability standard that ensures common operating systems and software packages can run seamlessly in ARM-based systems,” Mark Lohmeyer, Google Cloud’s VP for compute and AI/ML infrastructure, explained. “Through this collaboration, we’re accessing a broad ecosystem of cloud customers who have already deployed ARM-based workloads across hundreds of ISVs and open source projects.”

More later this year.


Software Development in Sri Lanka

Robotic Automations

Nvidia's next-gen Blackwell platform will come to Google Cloud in early 2025 | TechCrunch


It’s Google Cloud Next in Las Vegas this week, and that means it’s time for a bunch of new instance types and accelerators to hit the Google Cloud Platform. In addition to the new custom Arm-based Axion chips, most of this year’s announcements are about AI accelerators, whether built by Google or from Nvidia.

Only a few weeks ago, Nvidia announced its Blackwell platform. But don’t expect Google to offer those machines anytime soon. Support for the high-performance Nvidia HGX B200 for AI and HPC workloads and GB200 NBL72 for large language model (LLM) training will arrive in early 2025. One interesting nugget from Google’s announcement: The GB200 servers will be liquid-cooled.

This may sound like a bit of a premature announcement, but Nvidia said that its Blackwell chips won’t be publicly available until the last quarter of this year.

Image Credits: Frederic Lardinois/TechCrunch

Before Blackwell

For developers who need more power to train LLMs today, Google also announced the A3 Mega instance. This instance, which the company developed together with Nvidia, features the industry-standard H100 GPUs but combines them with a new networking system that can deliver up to twice the bandwidth per GPU.

Another new A3 instance is A3 confidential, which Google described as enabling customers to “better protect the confidentiality and integrity of sensitive data and AI workloads during training and inferencing.” The company has long offered confidential computing services that encrypt data in use, and here, once enabled, confidential computing will encrypt data transfers between Intel’s CPU and the Nvidia H100 GPU via protected PCIe. No code changes required, Google says. 

As for Google’s own chips, the company on Tuesday launched its Cloud TPU v5p processors — the most powerful of its homegrown AI accelerators yet — into general availability. These chips feature a 2x improvement in floating point operations per second and a 3x improvement in memory bandwidth speed.

Image Credits: Frederic Lardinois/TechCrunch

All of those fast chips need an underlying architecture that can keep up with them. So in addition to the new chips, Google also announced Tuesday new AI-optimized storage options. Hyperdisk ML, which is now in preview, is the company’s next-gen block storage service that can improve model load times by up to 3.7x, according to Google.

Google Cloud is also launching a number of more traditional instances, powered by Intel’s fourth- and fifth-generation Xeon processors. The new general-purpose C4 and N4 instances, for example, will feature the fifth-generation Emerald Rapids Xeons, with the C4 focused on performance and the N4 on price. The new C4 instances are now in private preview, and the N4 machines are generally available today.

Also new, but still in preview, are the C3 bare-metal machines, powered by older fourth-generation Intel Xeons, the X4 memory-optimized bare metal instances (also in preview) and the Z3, Google Cloud’s first storage-optimized virtual machine that promises to offer “the highest IOPS for storage optimized instances among leading clouds.”


Software Development in Sri Lanka

Robotic Automations

Google launches Code Assist, its latest challenger to GitHub's Copilot | TechCrunch


At its Cloud Next conference, Google on Tuesday unveiled Gemini Code Assist, its enterprise-focused AI code completion and assistance tool.

If this sounds familiar, that’s likely because Google previously offered a similar service under the now-defunct Duet AI branding. That one became generally available in late 2023, but even then, Google already hinted that it would move the service away from its Codey model to Gemini in the near future. Code Assist is both a rebrand of the older service as well as a major update.

Code Assist, which Google Cloud demoed at its 30,000-attendee conference in Las Vegas, will be available through plug-ins for popular editors like VS Code and JetBrains.

Even more so than the Duet AI version, Code Assist is also a direct competitor to GitHub’s Copilot Enterprise and not so much the basic version of Copilot. That’s because of a few Google-specific twists.

Among those is support for Gemini 1.5 Pro, which famously has a million-token context window, allowing Google’s tool to pull in a lot more context than its competitors. Google says this means more-accurate code suggestions, for example, but also the ability to reason over and change large chunks of code.

“This upgrade brings a massive 1 million-token context window, which is the largest in the industry. This allows customers to perform large-scale changes across your entire code base, enabling AI-assisted code transformations that were not possible before,” Brad Calder, Google’s VP and GM for its cloud platform and technical infrastructure, explained in a press conference ahead of Tuesday’s announcement.

Image Credits: Google

Like GitHub Enterprise, Code Assist can also be fine-tuned based on a company’s internal code base.

“Code customization using RAG with Gemini Code Assist significantly increased the quality of Gemini’s assistance for our developers in terms of code completion and generation,” said Kai Du, Director of Engineering and Head of Generative AI at Turing. “With code customization in place, we are expecting a big increase in the overall code-acceptance rate.”

This functionality is currently in preview.

Image Credits: Frederic Lardinois/TechCrunch

Another feature that makes Code Assist stand out is its ability to support codebases that sit on-premises, in GitLab, GitHub and Atlassian’s BitBucket, for example, as well as those that may be split between different services. That’s something Google’s most popular competitors in this space don’t currently offer.

Google is also partnering with a number of developer-centric companies to bring their knowledge bases to Gemini. Stack Overflow already announced its partnership with Google Cloud earlier this year. Datadog, Datastax, Elastic, HashiCorp, Neo4j, Pinecone, Redis, Singlestore and Snyk are now also partnering with Google through similar partnerships.

The real test, of course, is how developers will react to Code Assist and how useful its suggestions are to them. Google is making the right moves here by supporting a variety of code repositories and offering a massive context window, but if the latency is too high or the results simply aren’t that good, none of those features matter. And if it’s not significantly better than Copilot, which had quite a headstart, it may end up suffering the fate of AWS’ CodeWhisperer, which seems to have close to zero momentum.

It’s worth noting that in addition to Code Assist, Google today also announced the launch of CodeGemma, a new open model in its Gemma family that was fine-tuned for code generation and assistance. CodeGemma is now available through Vertex AI.

Image Credits: Frederic Lardinois/TechCrunch

Cloud Assist

In addition to Code Assist, Google also today announced Gemini Cloud Assist to help “cloud teams design, operate, and optimize their application lifecycle.” The tool can generate architecture configuration that are tailored to a company’s needs, for example, based on a description of the desired design outcome. It can also help diagnose issues and find their root causes, as well as optimize a company’s cloud usage to reduce cost or improve performance.

Cloud Assist will be available through a chat interface and embedded directly into a number of Google Cloud products.


Software Development in Sri Lanka

Robotic Automations

Google Workspace users will soon get voice prompting in Gmail and tabs in Docs | TechCrunch


Google continues to bring more AI-driven features to its Workspace productivity applications.

At its Cloud Next conference in Las Vegas, the company on Tuesday announced that Google Workspace subscribers will soon be able to use voice prompts to kick off the AI-based “Help me write” feature in Gmail while on the go, for example. In addition, Google is also launching a new feature in Gmail for Workspace that can instantly turn rough email drafts into a more polished email.

Image Credits: Google

These features will come to paying subscribers first. When asked about this in a press conference ahead of Tuesday’s announcements, Google’s Aparna Pappu noted that the company has “a long history of doing really useful, high-utility features with AI for all our users — including smart reply and smart compose. As we figure out how these work and get feedback from our users, we’ll consider expanding it to all our users.”

Workspace, which according to Google has about 3 billion users and over 10 million paying customers, was one of the first Google services to lean into the AI boom.

In addition to these new AI features, Google is adding a few other capabilities to the Workspace suite. These include notifications for Sheets, where the service will send out a customizable alert when a certain field changes, for example. In addition, Sheets will now feature a new set of templates to make getting started with a new spreadsheet easier.

And Docs, Google’s browser-based MS Word competitor, is getting support for tabs so “you can organize information in a single document instead of linking to multiple documents or searching through Drive to find what you’re looking for.” That’s a nifty feature and could be quite useful for workflows where you’d otherwise copy and paste a bunch of documents into one long one.  

Docs is also getting full-bleed cover images, and for those really large companies that use Workspace, Chat can now handle up to 500,000 members. Thanks to Google’s partnership with Mio, messaging interoperability with Slack and Teams is now an option, too.


Software Development in Sri Lanka

Robotic Automations

Google bets on partners to run their own sovereign Google Clouds | TechCrunch


Data sovereignty and residency laws have become commonplace in recent years. The major clouds, however, were always set up to enable the free movement of data between their various locations, so over the course of the last few years, all of the hyperscalers started looking into how they could offer sovereign clouds that can guarantee that government data, for example, never left a given country. AWS announced its European Sovereign Cloud last October. The Microsoft Azure Cloud for Sovereignty became generally available in December.

Google Cloud’s approach has been a bit different. Back in 2021, Google Cloud partnered with T-Systems to offer a sovereign cloud for Germany. A few weeks ago, it also announced a new partnership with World Wide Technology (WWT) to offer sovereign cloud solutions for government customers in the U.S.

Now Google is renewing its focus on data sovereignty. For the time being, though, it looks like its emphasis is on partnerships, not building its own sovereign clouds.

Google Cloud’s hybrid and on-premises story has changed quite a bit over the last few years. From the Cloud Services Platform to Anthos, GKE On-Prem and likely a few others that time has long forgotten, Google Cloud has aimed to offer a solution for companies that want to use its services and tooling but because of regulations, security, cost or paranoia, don’t want their workloads and data to sit in the Google cloud. Google’s latest effort in this space is branded Google Distributed Cloud (GDC), a fully managed software and hardware solution that can either be connected to the Google Cloud or be completely air-gapped from the internet.

Of course, this wouldn’t be 2024 if Google didn’t put an emphasis on AI in all of these efforts, too.

“Today, customers are looking for entirely new ways to process and analyze data, discover hidden insights, increase productivity and build entirely new applications — all with AI at the core,” said Vithal Shirodkar, VP/GM, Google Distributed Cloud and Geo Expansion, Google Cloud, in Tuesday’s announcement. “However, data sovereignty, regulatory compliance, and low-latency requirements can present a dilemma for organizations eager to adopt AI in the cloud. The need to keep sensitive data in certain locations, adhere to strict regulations, and ensure swift responsiveness can make it difficult to capitalize on the cloud’s inherent advantages of innovation, scalability, and cost-efficiency.”

At Cloud Next, Google Cloud’s annual developer conference, GDC is getting a slew of updates, including new security features (in partnership with Palo Alto Networks), support for the Apigee API management service and more. Developers can also now use a GDC Sandbox in Google Cloud to build and test applications without the need to work with the physical hardware. What’s maybe just as important as these new features is that GDC is now ISO27001 and SOC2 compliant.

On the hardware side, Google Cloud is introducing new AI servers for GDC. These are powered by Nvidia’s L4 Tensor Core GPUs and are now available in addition to the existing GDC AI-optimized servers with the high-powered Nvidia H100 GPUs.

Another interesting aspect to the GDC digital sovereignty story is that Google Cloud is emphasizing its partners, T-Systems, WWT and Clarence, which can deliver sovereign GDC-powered clouds on behalf of their clients.


Software Development in Sri Lanka

Robotic Automations

Google goes all in on generative AI at Google Cloud Next | TechCrunch


This week in Las Vegas, 30,000 folks came together to hear the latest and greatest from Google Cloud. What they heard was all generative AI, all the time. Google Cloud is first and foremost a cloud infrastructure and platform vendor. If you didn’t know that, you might have missed it in the onslaught of AI news.

Not to minimize what Google had on display, but much like Salesforce last year at its New York City traveling road show, the company failed to give all but a passing nod to its core business — except in the context of generative AI, of course.

Google announced a slew of AI enhancements designed to help customers take advantage of the Gemini large language model (LLM) and improve productivity across the platform. It’s a worthy goal, of course, and throughout the main keynote on Day 1 and the Developer Keynote the following day, Google peppered the announcements with a healthy number of demos to illustrate the power of these solutions.

But many seemed a little too simplistic, even taking into account they needed to be squeezed into a keynote with a limited amount of time. They relied mostly on examples inside the Google ecosystem, when almost every company has much of their data in repositories outside of Google.

Some of the examples actually felt like they could have been done without AI. During an e-commerce demo, for example, the presenter called the vendor to complete an online transaction. It was designed to show off the communications capabilities of a sales bot, but in reality, the step could have been easily completed by the buyer on the website.

That’s not to say that generative AI doesn’t have some powerful use cases, whether creating code, analyzing a corpus of content and being able to query it, or being able to ask questions of the log data to understand why a website went down. What’s more, the task and role-based agents the company introduced to help individual developers, creative folks, employees and others, have the potential to take advantage of generative AI in tangible ways.

But when it comes to building AI tools based on Google’s models, as opposed to consuming the ones Google and other vendors are building for its customers, I couldn’t help feeling that they were glossing over a lot of the obstacles that could stand in the way of a successful generative AI implementation. While they tried to make it sound easy, in reality, it’s a huge challenge to implement any advanced technology inside large organizations.

Big change ain’t easy

Much like other technological leaps over the last 15 years — whether mobile, cloud, containerization, marketing automation, you name it — it’s been delivered with lots of promises of potential gains. Yet these advancements each introduce their own level of complexity, and large companies move more cautiously than we imagine. AI feels like a much bigger lift than Google, or frankly any of the large vendors, is letting on.

What we’ve learned with these previous technology shifts is that they come with a lot of hype and lead to a ton of disillusionment. Even after a number of years, we’ve seen large companies that perhaps should be taking advantage of these advanced technologies still only dabbling or even sitting out altogether, years after they have been introduced.

There are lots of reasons companies may fail to take advantage of technological innovation, including organizational inertia; a brittle technology stack that makes it hard to adopt newer solutions; or a group of corporate naysayers shutting down even the most well-intentioned initiatives, whether legal, HR, IT or other groups that, for a variety of reasons, including internal politics, continue to just say no to substantive change.

Vineet Jain, CEO at Egnyte, a company that concentrates on storage, governance and security, sees two types of companies: those that have made a significant shift to the cloud already and that will have an easier time when it comes to adopting generative AI, and those that have been slow movers and will likely struggle.

He talks to plenty of companies that still have a majority of their tech on-prem and have a long way to go before they start thinking about how AI can help them. “We talk to many ‘late’ cloud adopters who have not started or are very early in their quest for digital transformation,” Jain told TechCrunch.

AI could force these companies to think hard about making a run at digital transformation, but they could struggle starting from so far behind, he said. “These companies will need to solve those problems first and then consume AI once they have a mature data security and governance model,” he said.

It was always the data

The big vendors like Google make implementing these solutions sound simple, but like all sophisticated technology, looking simple on the front end doesn’t necessarily mean it’s uncomplicated on the back end. As I heard often this week, when it comes to the data used to train Gemini and other large language models, it’s still a case of “garbage in, garbage out,” and that’s even more applicable when it comes to generative AI.

It starts with data. If you don’t have your data house in order, it’s going to be very difficult to get it into shape to train the LLMs on your use case. Kashif Rahamatullah, a Deloitte principal who is in charge of the Google Cloud practice at his firm, was mostly impressed by Google’s announcements this week, but still acknowledged that some companies that lack clean data will have problems implementing generative AI solutions. “These conversations can start with an AI conversation, but that quickly turns into: ‘I need to fix my data, and I need to get it clean, and I need to have it all in one place, or almost one place, before I start getting the true benefit out of generative AI,” Rahamatullah said.

From Google’s perspective, the company has built generative AI tools to more easily help data engineers build data pipelines to connect to data sources inside and outside of the Google ecosystem. “It’s really meant to speed up the data engineering teams, by automating many of the very labor-intensive tasks involved in moving data and getting it ready for these models,” Gerrit Kazmaier, vice president and general manager for database, data analytics and Looker at Google, told TechCrunch.

That should be helpful in connecting and cleaning data, especially in companies that are further along the digital transformation journey. But for those companies like the ones Jain referenced — those that haven’t taken meaningful steps toward digital transformation — it could present more difficulties, even with these tools Google has created.

All of that doesn’t even take into account that AI comes with its own set of challenges beyond pure implementation, whether it’s an app based on an existing model, or especially when trying to build a custom model, says Andy Thurai, an analyst at Constellation Research. “While implementing either solution, companies need to think about governance, liability, security, privacy, ethical and responsible use and compliance of such implementations,” Thurai said. And none of that is trivial.

Executives, IT pros, developers and others who went to GCN this week might have gone looking for what’s coming next from Google Cloud. But if they didn’t go looking for AI, or they are simply not ready as an organization, they may have come away from Sin City a little shell-shocked by Google’s full concentration on AI. It could be a long time before organizations lacking digital sophistication can take full advantage of these technologies, beyond the more-packaged solutions being offered by Google and other vendors.


Software Development in Sri Lanka

Back
WhatsApp
Messenger
Viber