Why AI needs a new approach to unstructured, fast-changing data: MongoDB Exec

MMS Founder
MMS RSS

Why AI needs a new approach to unstructured, fast-changing data: MongoDB Exec

<!—->


Loading...
Why AI needs a new approach to unstructured, fast-changing data: MongoDB Exec

As Indian enterprises push forward with digital transformation, data modernization has become a central focus. Companies are moving to cloud-native systems, using real-time data, and beginning to build AI-driven applications. 

MongoDB is working across the ecosystem, from startups and software vendors to large enterprises, to support this shift.  

In this conversation, Sachin Chawla, VP for India & ASEAN at MongoDB, talks about how Indian enterprises are approaching data transformation, the challenges they face, and how MongoDB is adapting its strategy to meet these changing needs. Edited Excerpts:  

What’s fundamentally changing in how Indian enterprises approach data and digital transformation, and how is that shaping your strategy in the region? 

India is going through a significant phase in technology. While we’ll get to the challenges shortly, it’s worth highlighting the strong momentum we’re seeing in modern application development across the ecosystem. 

This includes the startup space, where innovation is active, from early-stage companies like RFPIO to large-scale players like Zepto and Zomato. There’s also a robust ISV ecosystem here, which operates quite differently compared to markets like the US or EMEA. For instance, many Indian banks rely heavily on ISVs, often referred to as “bank tech”, for their software needs. Companies like CRMnext, Intellect AI, and Lentra are key examples, and many of them use MongoDB. 

Beyond startups and ISVs, significant digital transformation is happening in large enterprises as well. Take Tata, for example, the Tata Neu app runs on MongoDB. So overall, there’s a lot of progress and activity in the ecosystem. 

Now, on to the challenges. Broadly, they’re similar to those faced globally and can be grouped into three areas, first, improving developer productivity. Every organization is focused on how to help developers move faster and be more efficiently. Second, building modern AI applications. There’s growing pressure, both from the ecosystem and from leadership, to deliver AI-driven solutions. 

Third, modernizing legacy applications. Many existing systems were built over years and aren’t designed to meet today’s demands. Users expect immediate, responsive digital experiences, and older systems can’t keep up. These are the key priorities: boosting developer productivity, adopting AI, and modernizing legacy systems. 

What are the biggest misconceptions Indian enterprises still have about modern databases, and how do these hold back their digital transformation? 

First, some organizations treat modernization as simply moving their existing systems from on-premises to the cloud. But lifting and shifting the stack without changing the application, the underlying infrastructure, or the database doesn’t actually modernize anything. It’s still the same application, just running in a different environment, and it won’t deliver new value just by being in the cloud. 

Second, there’s the idea of using the best purpose-built database for each use case. In practice, this often leads to an overload of different databases within one organization. While each one might be optimal for its specific function, managing a large variety becomes a challenge. You end up spending more time and resources maintaining the system than actually innovating with it. 

Third, when it comes to AI, many organizations lack a clear strategy. They often start building without clearly defined use cases or objectives, just reacting to pressure to “do something with AI.” Without a focused plan, it’s hard to deliver meaningful outcomes. 

Which industries in India are making the boldest or most unexpected moves in digital transformation or data modernization, and why? 

Every sector is adopting AI in its own way. Tech startups and digital-native companies tend to make faster, more visible moves, but even large enterprises are accelerating their efforts. 

For example, we recently partnered with Intellect AI, an independent software vendor serving the global banking sector. They aimed to lead the way in building a multi-agent platform for banking clients to automate and augment operations in areas like compliance and risk, critical functions for many institutions. 

We helped them develop this platform using MongoDB and MongoDB’s AI vector search. The result is called Purple Fabric, and it’s publicly available. This platform is now driving automation and augmentation in compliance and risk management. 

One of their major clients is a sovereign fund managing $1.5 trillion in assets across around 9,000 companies. Using Purple Fabric, they automated ASC compliance processes with up to 90% accuracy. 

This example shows that while enterprises may seem slower, companies like Intellect AI are enabling them to move quickly by building powerful tools tailored for complex environments. 

What recurring data architecture issues do you see in enterprise AI projects, and how does your company help address them? 

When you look at AI applications, it’s important to understand that the data used to build them is mostly unstructured. This data is often messy, comes from various sources, and appears in different formats such as video, audio, and text. Much of it is interconnected, and the overall volume is massive. Additionally, the data changes rapidly. 

These are the three core characteristics of AI data: it’s unstructured, high in volume, and constantly evolving. As a result, if you look at an AI application a year after it’s built without any updates, it’s likely already outdated. Continuous updates are essential. 

MongoDB stores data in a document format, unlike traditional systems that use rows and columns. Trying to store large volumes of unstructured and fast-changing data in a tabular format becomes unmanageable. You’d end up with thousands of tables, all linked in complex ways. Any change in one table could affect hundreds or thousands of others, making maintenance difficult. 

This is why many modern applications are built on MongoDB rather than on legacy systems. For example, Intellect AI uses MongoDB, as does DarwinBox, which uses AI to power talent search queries like finding the percentage of top-rated employees. Previously, this kind of semantic search would take much longer. 

Another example is Ubuy, an e-commerce platform with around 300 million products. They switched from a SQL database to MongoDB with vector search. Search times went from several seconds to milliseconds, enabling efficient semantic search. 

RFP.io is another case. It uses vector search to process and understand RFP documents, identifying which sections relate to topics like security or disaster recovery. This simplifies the process of responding to RFPs. 

As enterprises shift to unstructured data, vector search, and real-time AI, how is MongoDB adapting, and what key industry gaps still remain? 

The first step is collecting and using data in real time. For that, you need the right database. A document model is a natural fit for the scale and structure of this data. 

Once you have the data, the next step is using it effectively. That starts with full-text search, similar to how you search on Google. Most applications today rely on this kind of search functionality. 

But if you’re building AI applications, you also need to vectorize the data. Without vectorization, you can’t perform semantic searches or build meaningful AI features. 

At this point, companies usually face a choice. They often have data spread across multiple databases. To enable full-text search, they might add a solution like Elasticsearch. For semantic search, they bring in a vector database such as Pinecone. If they want to train or fine-tune models using internal data, they also need an embedding model. So now they’re managing a database, a full-text search engine, a vector search system, and an embedding model, each a separate component. 

The integration work required to get all these systems to operate together can consume a large amount of development and management time, pulling focus away from innovation. 

In contrast, our platform simplifies this. It uses a single document database to store all types of data. It includes Atlas Search for full-text search, built-in vector search, and now, with our acquisition of Voyage, integrated embedding capabilities. You don’t need separate systems for each function. 

With everything in one place, there’s no need for complex integration. You can run full-text and semantic (hybrid) search from the same platform. This reduces cost, simplifies management, and frees up time for innovation. Customers tell us this is the biggest advantage—they don’t need to stitch multiple tools together, which can be very hard to manage. 

What’s next for MongoDB in India, are you focusing on AI, edge deployments, cloud-native capabilities, or something else? 

Our focus will remain on three main areas. First, we’ll continue working with developers to help them improve their productivity. Second, we’ll collaborate across the ecosystem and with enterprises to build modern applications. Third, and most significantly, we’ll support large enterprises as they modernize their applications, whether by creating new ones or upgrading legacy systems. This includes helping them reduce technical debt, move away from outdated applications and databases like Oracle and SQL, and transition to more modern architectures that align with their goals. These are our three key priorities. 

Where do you see AI heading in the data modernization space over the next three to five years? 

In my view, we’re at a stage similar to the 1960s when computers and operating systems were just emerging. I see large language models (LLMs) as the new operating systems. We’re in the early phase, and what comes next is the development of applications built on top of them. As this evolves, more advanced and diverse applications will emerge. 

Building applications is becoming much easier. For example, there’s a concept called white coding, where even young children, eight or nine years old—can create apps. If a computer can guide you step by step, almost anyone can build one. That’s where we’re heading: a world where millions of applications can be developed quickly and easily. 

We see ourselves as a natural platform for these applications because we make it simple to store data. So, over the next few years, we expect a surge in development activity. A lot is going to change, and I think we’ll all be surprised by just how much. 

Loading...


Next Article

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.

Google DeepMind Unveils AlphaGenome: A Unified AI Model for High-Resolution Genome Interpretation

MMS Founder
MMS Robert Krzaczynski

Google DeepMind has announced the release of AlphaGenome, a new AI model designed to predict how genetic variants affect gene regulation across the entire genome. It represents a significant advancement in computational genomics by integrating long-range sequence context with base-pair resolution in a single, general-purpose architecture.

AlphaGenome processes up to 1 million base-pairs of DNA at once and outputs high-resolution predictions across thousands of molecular modalities, including gene expression, chromatin accessibility, transcription start sites, RNA splicing, and protein binding. It allows researchers to evaluate the effects of both common and rare variants, not just in protein-coding regions, but in the far more complex non-coding regulatory regions that constitute 98% of the human genome.

Technically, AlphaGenome combines convolutional neural networks (CNNs) to detect local sequence motifs and transformers to model long-range interactions, all trained on rich multi-omic datasets from ENCODE, GTEx, 4D Nucleome, and FANTOM5. The architecture achieves state-of-the-art performance across a broad range of genomic benchmarks, outperforming task-specific models in 24 out of 26 evaluations of variant effect prediction.

A notable innovation is AlphaGenome’s ability to directly model RNA splice junctions, a feature crucial for understanding many genetic diseases caused by splicing errors. The model can also contrast mutated and reference sequences to quantify the regulatory impact of variants across tissues and cell types — a key capability for studying disease-associated loci and interpreting genome-wide association studies (GWAS).

Training efficiency was also improved: a full AlphaGenome model was trained in just four hours on TPUs, using half the compute budget of DeepMind’s earlier Enformer model, thanks to optimized architecture and data pipelines.

The model is now available via the AlphaGenome API for non-commercial research use, enabling scientists to generate functional hypotheses at scale without needing to combine disparate tools or models. DeepMind has indicated plans for further extension to new species, tasks, and fine-tuned clinical applications.

This release also aligns with a broader conversation around the interpretability and emotional context of AI in medicine. As Graevka Suvorov, an AI alignment researcher, commented:

The true frontier for MedGemma isn’t just diagnostic accuracy, but the informational and psychological state it creates in the patient. A diagnosis without context is a data point that can create fear. A diagnosis delivered with clarity is the first step to healing. An AI with a true ‘informational bedside manner’—one that understands it’s not just treating an image, but a person’s entire reality—is the next real leap in AGI.

AlphaGenome pushes the field closer to that vision, enabling deeper, more accurate interpretations of the genome and offering a unified model for understanding biology at the sequence level.

About the Author

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.

MongoDB’s Strategic Pivot: Securing Its Future in High-Security Cloud Databases – AInvest

MMS Founder
MMS RSS

MongoDB, Inc. (NASDAQ: MDB) has long been a leader in NoSQL databases, but its recent strategic moves—securing inclusion in the Russell Midcap Value Index and pursuing FedRAMP High authorization—signal a bold pivot toward positioning itself as a high-security cloud database provider for government and regulated industries. This repositioning could unlock significant growth opportunities while attracting new investors.

The Russell Midcap Value Inclusion: A Strategic Rebranding

MongoDB’s addition to the Russell Midcap Value Index in early 2024 marks a pivotal shift in its valuation narrative. While MDB has historically been classified as a growth stock (its price-to-sales ratio remains elevated at ~10x), its inclusion in a value-oriented index reflects a maturing business model and improving profitability.

This reclassification is no accident. MongoDB has prioritized margin expansion and recurring revenue streams, with its flagship Atlas cloud service now contributing 70% of total revenue (up from 66% in 2024). The Russell Midcap Value Index inclusion could attract investors seeking stable, cash-flow positive companies in the tech sector—a demographic MDB has historically struggled to engage.

FedRAMP High: Unlocking the $100B Government Cloud Market

MongoDB’s pursuit of FedRAMP High and Impact Level 5 (IL5) certifications by 2025 is its most critical strategic move. These certifications will enable MongoDB Atlas for Government to handle highly sensitive data, including national security and health records, which are currently off-limits to its cloud platform.

The stakes are enormous: U.S. federal cloud spending is projected to hit $100 billion by 2027, with security-conscious agencies favoring providers that meet the strictest compliance standards. While MongoDB currently holds FedRAMP Moderate authorization, the FedRAMP High upgrade—subject to 421 stringent security controls—will open access to lucrative contracts with defense, intelligence, and healthcare agencies.

A Case Study in Success: The Utah Migration

MongoDB’s partnership with the State of Utah offers a blueprint for its government strategy. By migrating its benefits eligibility system to Atlas, Utah reduced disaster recovery time from 58 hours to 5 minutes, while cutting costs and improving speed. This win highlights Atlas’s ability to modernize legacy systems securely, a key selling point for agencies wary of cloud adoption.

Financials Support the Shift to Security

MongoDB’s financials back its strategic pivot:
Q1 2025 revenue grew 22% YoY to $450.6 million, driven by 32% growth in Atlas revenue.
Customer count rose to 49,200, with 73% of $1M+ customers increasing spend.
Margin expansion: Gross margins improved to 68% in Q1 2025, up from 65% in 2024.

These metrics suggest MongoDB is executing its shift toward high-margin, subscription-based cloud services while scaling its salesforce to target regulated sectors.

Risks and Considerations

  • Competition: AWS, Microsoft, and Snowflake are aggressively targeting the government cloud market.
  • Certification Delays: FedRAMP High and IL5 approvals are pending, and delays could push revenue growth below expectations.
  • Valuation: MDB’s stock trades at a premium relative to peers (e.g., Snowflake’s P/S of ~3x).

Investment Thesis: A Buy with Long-Term Upside

MongoDB’s strategic moves—Russell Midcap Value inclusion and FedRAMP High pursuit—position it to capitalize on a $100B+ addressable market in secure cloud databases. While short-term risks exist, the long-term opportunity for MDB to dominate regulated sectors justifies its valuation.

Buy recommendation: With a $430 price target from Citigroup (108% upside from current levels) and strong hedge fund support, MDB is a speculative but high-reward play for investors willing to bet on its security-driven growth.

Final Take

MongoDB’s pivot to high-security cloud databases is more than a rebrand—it’s a calculated move to tap into one of the fastest-growing segments of the tech industry. If it secures FedRAMP High by 2025, MDB could emerge as a must-have partner for governments and enterprises, justifying its premium valuation. For investors, this is a story worth watching closely.

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.

MongoDB Announces Commitment to Achieve FedRAMP High and Impact Level 5 Authorizations

MMS Founder
MMS RSS

MongoDB, Inc. announced its commitment to pursuing Federal Risk and Authorization Management Program(FedRAMP) High and Impact Level 5(IL5) authorizations for MongoDB Atlas for Government workloads, which will expand its eligibility to manage unclassified, yet highly sensitive, U.S. public sector data. 

With FedRAMP High authorization, even the most critical government agencies looking to adopt cloud and AI technologies—and to modernize aging, inefficient legacy databases—can rely on MongoDB Atlas for Government for secure, fully managed workloads.

MongoDB Atlas for Government already provides a flexible way for the U.S. public sector to deploy, run, and scale modern applications in the cloud within a dedicated environment built for FedRAMP Moderate workloads. 

Achieving FedRAMP High and IL5 will allow MongoDB Atlas for Government’s secure, reliable, and high-performance modern database solutions to be used to manage high-impact data, such as in emergency services, law enforcement systems, financial systems, health systems, and any other system where loss of confidentiality, integrity, or availability could have a severe or catastrophic adverse effect on organizational operations, organizational assets, or individuals.

“The federal agencies that manage highly sensitive data involving the protection of life and financial ruin should be using the latest, fastest, and best database technology available,” said Benjamin Cefalo, Senior Vice President of Product Management at MongoDB. 

“With FedRAMP High and IL5 authorizations for MongoDB Atlas for Government workloads, they will be able to take advantage of MongoDB’s industry-leading and proprietary Queryable Encryption, multi-cloud flexibility and resilience, high availability with automated backup, data recovery options, and on-demand scaling, and native vector search to facilitate building AI applications.”

MongoDB Atlas for Government already helps hundreds of public sector agencies nationwide develop secure, modern, and scalable solutions. An integral feature of MongoDB Atlas for Government is MongoDB Queryable Encryption. 

This industry-first, in-use encryption technology enables organizations to encrypt sensitive data that helps organizations protect sensitive data when it is queried and in use on Atlas for Government. 

With Queryable Encryption, sensitive data remains protected throughout its lifecycle, whether it is in-transit, at-rest, in-use, and in logs and backups. It is only ever decrypted on the client-side.

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.

MongoDB Announces Commitment To Pursuing FedRAMP High And Impact Level 5 …

MMS Founder
MMS RSS

<!–>

With upgraded authorization, soon federal agencies with security requirements at every level will be able to use MongoDB to deploy, run, and scale modern applications in the cloud

NEW YORK, June 30, 2025 /PRNewswire/ – MongoDB, Inc. (NASDAQ:MDB) today announced its commitment to pursuing Federal Risk and Authorization Management Program (FedRAMP) High and Impact Level 5 (IL5) authorizations for MongoDB Atlas for Government workloads, which will expand its eligibility to manage unclassified, yet highly sensitive, U.S. public sector data. With FedRAMP High authorization, even the most critical government agencies looking to adopt cloud and AI technologies—and to modernize aging, inefficient legacy databases—can rely on MongoDB Atlas for Government for secure, fully managed workloads.

–>

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.