Storage and Data Protection News for the Week of June 20; Updates from Cohesity, Pure …

MMS Founder
MMS RSS

Posted on mongodb google news. Visit mongodb google news

Solutions Review Executive Editor Tim King curated this list of notable storage and data protection news for the week of June 20, 2025.

Keeping tabs on all the most relevant storage and data protection news can be a time-consuming task. As a result, our editorial team aims to provide a summary of the top headlines from the last week in this space. Solutions Review editors will curate vendor product news, mergers and acquisitions, venture capital funding, talent acquisition, and other noteworthy storage and data protection news items.

For early access to all the expert insights published on Solutions Review, join Insight Jam, a community dedicated to enabling the human conversation on AI.

Top Storage and Data Protection News for the Week of June 20, 2025


Cohesity Boosts Resilience for Mission-Critical MongoDB Workloads

Cohesity has announced a deeper integration with MongoDB, delivering advanced backup and recovery capabilities tailored for large, mission-critical datasets. The new API-based solution enhances performance, scalability, and security—supporting the stringent requirements of global enterprises in banking, finance, and beyond.

Read more: prnewswire.com/apac/news-releases/cohesity-strengthens-resilience-of-large-mission-critical-mongodb-workloads-302485292.html

CTERA First in Hybrid Cloud Storage to Support Model Context Protocol (MCP)

CTERA has become the first hybrid cloud storage provider to support the Model Context Protocol (MCP), enabling seamless integration with next-generation AI and machine learning workflows. This innovation allows enterprises to securely manage and access contextual data across distributed environments, paving the way for more intelligent, scalable AI deployments.

Read more: globenewswire.com/fr/news-release/2025/06/17/3100715/0/en/CTERA-Becomes-First-in-Hybrid-Cloud-Storage-to-Support-the-Model-Context-Protocol-MCP.html

DoControl Launches Dot: The First AI-Powered SaaS Data Security Assistant

DoControl has introduced Dot, the industry’s first AI-powered SaaS Data Security Assistant designed to revolutionize how security teams manage and protect SaaS environments. Leveraging a deeply contextualized data infrastructure, Dot enables natural-language interaction with complex security data, transforming fragmented, manual processes into simple, conversational workflows.

Read more: prnewswire.com/news-releases/docontrol-launches-first-ever-ai-powered-saas-data-security-assistant-dot-302484628.html

Study Finds 72 Percent of Enterprises Plan to Expand AI Use Despite Data Privacy and Ethics Concerns

A recent survey reveals that while 72 percent of enterprises intend to broaden their AI adoption, many remain concerned about data privacy, bias, and ethical risks. Tech leaders prioritize expanding AI capabilities but emphasize the need for robust governance frameworks to mitigate these challenges and ensure responsible AI deployment across industries.

Read more: finance.yahoo.com/news/study-finds-72-enterprises-plan-130000001.html

IBM Introduces Industry-First Software for Unified Agentic Governance and Security

IBM has launched the industry’s first software platform designed to unify agentic governance and security for enterprise AI. The new solution provides organizations with centralized oversight and control of AI agents, ensuring compliance, risk management, and operational transparency as businesses deploy increasingly autonomous AI systems across their operations.

Read more: newsroom.ibm.com/2025-06-18-ibm-introduces-industry-first-software-to-unify-agentic-governance-and-security

NTT DATA Report Reveals C-Suite Divide on GenAI Adoption and Security

A new global report from NTT DATA highlights a significant misalignment within the C-suite regarding generative AI (GenAI) adoption. While 99 percent of executives plan further GenAI investments and 67 percent of CEOs are making significant commitments, nearly half of CISOs express concerns about security gaps and unclear internal guidelines.

Read more: https://www.ndtvprofit.com/technology/c-suite-misalignment-over-gen-ai-adoption-shows-ntt-data-research

Nutanix Report: Public Sector Rapidly Adopting Generative AI, But Faces Security Hurdles

A new Nutanix study reveals that 83 percent of public sector organizations have a generative AI strategy, with most already implementing or preparing to deploy AI solutions. While GenAI is expected to boost productivity and automation, 76 percent of IT leaders say their infrastructure needs significant upgrades to support these advances. Security and privacy remain top concerns, with 92 percent of leaders calling for stronger protections as AI becomes integral to public sector operations.

Read more: nutanix.com/press-releases/2025/nutanix-study-finds-public-sector-embraces-generative-ai

Pure Storage Introduces the Enterprise Data Cloud

Pure Storage has unveiled its Enterprise Data Cloud, a new platform designed to unify data management, storage, and analytics for modern enterprises. The solution promises seamless data mobility, robust security, and simplified operations—helping organizations harness the full value of their data across on-premises and cloud environments.

Read more: purestorage.com/company/newsroom/press-releases/pure-storage-introduces-the-enterprise-data-cloud.html

Qumulo Stratus: Cryptographically Isolated Edge-to-Core-to-Cloud Data Platform

Qumulo has announced Stratus, a multi-tenant data platform that delivers cryptographically isolated data management from edge to core to cloud. Stratus is designed to empower organizations with secure, scalable, and resilient data storage, supporting modern workloads and regulatory requirements across distributed environments.

Read more: businesswire.com/news/home/20250617477912/en/Qumulo-Stratus-Cryptographically-Isolated-Edge-to-CoretoCloud-MultiTenant-Data-Platform

SingleStore Unveils Enterprise AI Platform for Real-Time, Serverless Workflows

SingleStore has launched a major update to its enterprise AI platform, introducing real-time, serverless functions designed to streamline development and deliver ultimate flexibility for AI-powered applications. The new release focuses on enhancing developer productivity and enabling organizations to build, deploy, and scale AI solutions faster and more efficiently.

Read more: aithority.com/machine-learning/singlestore-drops-an-enterprise-ai-glow-up-built-for-real-time-serverless-functions-and-ultimate-dev-flow/

StorONE Launches OneAI to Revolutionize Storage Intelligence

StorONE has introduced OneAI, a new platform that brings advanced artificial intelligence to storage management. OneAI is designed to optimize performance, automate routine tasks, and provide predictive analytics, enabling organizations to maximize the value and efficiency of their storage infrastructure.

Read more: storone.com/news-and-events/storone-unveils-oneai/

Expert Insights

Watch this space each week as our editors will share upcoming events, new thought leadership, and the best resources from Insight Jam, Solutions Review’s enterprise tech community where the human conversation around AI is happening. The goal? To help you gain a forward-thinking analysis and remain on-trend through expert advice, best practices, predictions, and vendor-neutral software evaluation tools.

Take the Tech Leader Survey – Spring 2025 Now

In partnership with Skiilify Co-Founder and distinguished Northeastern University Professor Paula Caligiuri, PhD, we’ve just launched our latest enterprise tech leader Survey to uncover howthought leaders are thinking about disruption in this AI moment.

Take survey

NEW by SR Expert at Insight Jam Paula Caligiuri, PhD.: How Humility Supercharges Technical Talent

According to our April 2025 report, 81 percent of tech professionals agree that humility—defined as actively seeking and using feedback—is critical for growth. But nearly half (46 percent) say the feedback they receive is too vague or unhelpful and not actionable. Which means even the most open-minded, competent professionals are flying without enough external input to keep growing.

Read on Solutions Review

ICYMI: Mini Jam Q2, 2025 – AI Agents of Change: Watch All Sessions On-Demand

Centered on the theme ‘AI Agent of Change’, this quarterly Insight Jam virtual event explores the emerging role of AI agents in reshaping the Future of  Work through Automation, driving innovation, and facilitating adaptation to the new reality of this AI moment. Be sure to register free for InsightJam.com here to watch all the sessions live or on-demand.

Read on Solutions Review

For consideration in future data protection news roundups, send your announcements to the editor: tking@solutionsreview.com.

Article originally posted on mongodb google news. Visit mongodb google news

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


MongoDB, Inc. (NASDAQ:MDB) Stock Position Increased by Fifth Third Bancorp

MMS Founder
MMS RSS

Posted on mongodb google news. Visit mongodb google news

Fifth Third Bancorp boosted its position in MongoDB, Inc. (NASDAQ:MDBFree Report) by 15.9% during the 1st quarter, according to the company in its most recent filing with the Securities & Exchange Commission. The fund owned 569 shares of the company’s stock after purchasing an additional 78 shares during the quarter. Fifth Third Bancorp’s holdings in MongoDB were worth $100,000 at the end of the most recent quarter.

Several other hedge funds have also recently bought and sold shares of the company. Assenagon Asset Management S.A. raised its position in shares of MongoDB by 24.4% during the 1st quarter. Assenagon Asset Management S.A. now owns 369,313 shares of the company’s stock worth $64,778,000 after buying an additional 72,424 shares in the last quarter. Handelsbanken Fonder AB raised its position in shares of MongoDB by 0.4% during the 1st quarter. Handelsbanken Fonder AB now owns 14,816 shares of the company’s stock worth $2,599,000 after buying an additional 65 shares in the last quarter. SG Americas Securities LLC raised its position in shares of MongoDB by 1,230.3% during the 1st quarter. SG Americas Securities LLC now owns 24,997 shares of the company’s stock worth $4,384,000 after buying an additional 23,118 shares in the last quarter. Farther Finance Advisors LLC raised its position in shares of MongoDB by 57.2% during the 1st quarter. Farther Finance Advisors LLC now owns 1,242 shares of the company’s stock worth $219,000 after buying an additional 452 shares in the last quarter. Finally, Park Avenue Securities LLC raised its position in shares of MongoDB by 52.6% during the 1st quarter. Park Avenue Securities LLC now owns 2,630 shares of the company’s stock worth $461,000 after buying an additional 907 shares in the last quarter. 89.29% of the stock is currently owned by institutional investors.

Analyst Ratings Changes

A number of analysts have commented on the stock. Rosenblatt Securities dropped their price target on shares of MongoDB from $305.00 to $290.00 and set a “buy” rating on the stock in a research report on Thursday, June 5th. Needham & Company LLC restated a “buy” rating and set a $270.00 price target on shares of MongoDB in a research report on Thursday, June 5th. Macquarie restated a “neutral” rating and set a $230.00 price target (up from $215.00) on shares of MongoDB in a research report on Friday, June 6th. Daiwa Capital Markets assumed coverage on shares of MongoDB in a report on Tuesday, April 1st. They issued an “outperform” rating and a $202.00 target price on the stock. Finally, Citigroup dropped their target price on shares of MongoDB from $430.00 to $330.00 and set a “buy” rating on the stock in a report on Tuesday, April 1st. Eight investment analysts have rated the stock with a hold rating, twenty-four have issued a buy rating and one has issued a strong buy rating to the stock. According to MarketBeat.com, the company has a consensus rating of “Moderate Buy” and an average price target of $282.47.

<!—->

Read Our Latest Analysis on MongoDB

MongoDB Stock Down 0.7%

MDB stock opened at $204.15 on Friday. The company has a market capitalization of $16.68 billion, a price-to-earnings ratio of -179.08 and a beta of 1.39. MongoDB, Inc. has a 12 month low of $140.78 and a 12 month high of $370.00. The firm has a 50 day moving average of $184.32 and a two-hundred day moving average of $223.66.

MongoDB (NASDAQ:MDBGet Free Report) last posted its earnings results on Wednesday, June 4th. The company reported $1.00 EPS for the quarter, topping the consensus estimate of $0.65 by $0.35. The company had revenue of $549.01 million during the quarter, compared to the consensus estimate of $527.49 million. MongoDB had a negative net margin of 4.09% and a negative return on equity of 3.16%. The company’s revenue for the quarter was up 21.8% compared to the same quarter last year. During the same period in the previous year, the firm posted $0.51 earnings per share. On average, research analysts expect that MongoDB, Inc. will post -1.78 EPS for the current fiscal year.

Insider Buying and Selling

In related news, CEO Dev Ittycheria sold 25,005 shares of the company’s stock in a transaction on Thursday, June 5th. The shares were sold at an average price of $234.00, for a total transaction of $5,851,170.00. Following the completion of the sale, the chief executive officer now directly owns 256,974 shares of the company’s stock, valued at approximately $60,131,916. This represents a 8.87% decrease in their position. The transaction was disclosed in a filing with the Securities & Exchange Commission, which is accessible through this hyperlink. Also, insider Cedric Pech sold 1,690 shares of the company’s stock in a transaction on Wednesday, April 2nd. The shares were sold at an average price of $173.26, for a total transaction of $292,809.40. Following the transaction, the insider now owns 57,634 shares of the company’s stock, valued at approximately $9,985,666.84. The trade was a 2.85% decrease in their ownership of the stock. The disclosure for this sale can be found here. In the last three months, insiders sold 49,208 shares of company stock valued at $10,167,739. Company insiders own 3.10% of the company’s stock.

MongoDB Profile

(Free Report)

MongoDB, Inc, together with its subsidiaries, provides general purpose database platform worldwide. The company provides MongoDB Atlas, a hosted multi-cloud database-as-a-service solution; MongoDB Enterprise Advanced, a commercial database server for enterprise customers to run in the cloud, on-premises, or in a hybrid environment; and Community Server, a free-to-download version of its database, which includes the functionality that developers need to get started with MongoDB.

Featured Stories

Institutional Ownership by Quarter for MongoDB (NASDAQ:MDB)



Receive News & Ratings for MongoDB Daily – Enter your email address below to receive a concise daily summary of the latest news and analysts’ ratings for MongoDB and related companies with MarketBeat.com’s FREE daily email newsletter.

Article originally posted on mongodb google news. Visit mongodb google news

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


MongoDB Launches an Open Source Real-Time Secret Scanner – It’s FOSS News

MMS Founder
MMS RSS

Posted on mongodb google news. Visit mongodb google news

Accidentally exposing secrets like API keys, tokens, or credentials in your code opens the door for threat actors to exploit your systems. Such attackers don’t stop at one breach; they automate their attacks, move fast, and can potentially compromise entire infrastructure within minutes.

To tackle such scenarios, MongoDB has come up with an open source solution called “Kingfisher“.

What’s Happening: Launched as an open source tool for detecting secrets in code, file systems, and Git history, Kingfisher was born out of MongoDB’s need for a fast, reliable way to identify exposed credentials and prevent security risks before they spiral out of control.

The tool doesn’t just stop there; it can also validate any secrets it finds, as long as they are from supported services, so developers know which keys are still active and risky.

MongoDB has been using Kingfisher internally throughout its development and deployment processes, helping them detect and fix exposed secrets early.

What to Expect: As for how it works, Kingfisher scans code, files, and Git history using various techniques like entropy analysis, real-time validation, pattern matching, and source code parsing for or accurate detection of exposed secrets.

It’s written in Rust and has many handy features like multi-language source parsing with Tree-sitter, high-speed regex matching with Hyperscan, extensible rulesets, cross-platform support, and over 700 built-in detection rules that cover a wide range of cloud services and secret types.

All of this runs on the user’s own systems or infrastructure, ensuring no sensitive data is sent to third-party servers, and there’s cross-platform support for Linux, Windows, and macOS. Using Kingfisher also helps security teams stay aligned with SLSA compliance standards.

If you are up for a longer read, then MongoDB has published a detailed blog post explaining how they built Kingfisher.

It’s FOSS turns 13! 13 years of helping people use Linux ❤️

And we need your help to go on for 13 more years. Support us with a Plus membership and enjoy an ad-free reading experience and get a Linux eBook for free.

To celebrate 13 years of It’s FOSS, we have a lifetime membership option with reduced pricing of just $76. This is valid until 25th June only.

If you ever wanted to appreciate our work with Plus membership but didn’t like the recurring subscription, this is your chance 😃


Get Lifetime Membership of It’s FOSS

Article originally posted on mongodb google news. Visit mongodb google news

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


Intelligent Cloud Efficiency: How Hari Babu Dama’s AI-Powered Optimization Model … – Outlook India

MMS Founder
MMS RSS

Posted on mongodb google news. Visit mongodb google news

As cloud-native digital operations become essential for enterprises, one critical question dominates boardrooms and DevOps pipelines alike: How can we reduce cloud database costs without compromising performance?

In an era of exponential data growth, the financial burden of managing transactional and analytical workloads in the cloud has become unsustainable. Organizations are struggling to manage rising costs from underutilized instances, suboptimal configurations, and lack of Observability into their workload patterns. However, the work of Haribabu, a cloud optimization professional and innovator in data efficiency, is helping to address these challenges.

In his recent study published in the International Journal of Innovative Research in Science, Engineering and Technology (IJIRSET), titled “Cloud Cost Optimization for Database Workloads: Real-World Savings Using Utilization Analytics”, Haribabu introduces an intelligent framework that leverages AI-based utilization tracking to reduce database costs by up to 47% in real-world cloud deployments.

“Optimization isn’t just about rightsizing resources,” Haribabu notes. “It’s about aligning computing power with intent, and making cost-efficiency a design principle, not an afterthought.”

From Over provisioning to AI-Led Utilization Intelligence

Haribabu’s framework introduces a data-driven utilization analytics engine that maps real-time database usage patterns across platforms like Amazon RDS, Azure SQL, Google Cloud SQL, and MongoDB Atlas.

Unlike conventional monitoring tools that capture static snapshots, his solution performs continuous workload profiling, capturing IOPS, memory bursts, connection counts, query types, and CPU spikes then correlates these with historical trends and business-critical SLAs.

Key findings from production deployments reveal:

  • 47.2% reduction in monthly cloud database costs

  • 28% improvement in query execution latency due to instance reallocation

  • Zero downtime optimization using phased migration

This approach aim to reduce both over provisioning (paying for unused capacity) and under provisioning (leading to slowdowns and outages). Instead, database instances are dynamically adjusted based on actual workload behavior—ensuring the perfect balance between performance and cost.

Multi-Cloud and Multi-Model Optimization

Today’s enterprises often operate in multi-cloud and multi-database environments—running OLTP workloads on MySQL in AWS, reporting workloads on Azure Synapse, and real-time analytics on NoSQL stores like Cassandra or DynamoDB.

Haribabu’s optimization model supports this diversity with a cloud-agnostic optimization engine. The platform interfaces with all major cloud service providers and supports multiple storage engines—relational, document, graph, time-series, and columnar.

It detects inefficiencies such as:

By recommending low-impact configuration changes like converting standard SSD to magnetic storage for non-critical backups or decommissioning underutilized indexes—The system identified opportunities for significant cost savings while maintaining data integrity.

Predictive Scaling and Cost Forecasting

A key component of Haribabu’s system is its AI-powered predictive analytics engine. Trained on workload patterns and calendar-based trends (quarter-ends, sales spikes, batch ETLs), the engine forecasts capacity needs days or even weeks in advance.

For instance, one use case involved a fintech company that faced high database latency every Monday due to weekly settlement jobs. Haribabu’s solution flagged this spike proactively, scaled up the instance type for just six hours, and then automatically scaled back achieving 100% availability and saving over $120,000 annually.

“Predictive scaling transforms DevOps from a reactive firefighting role into a strategic forecasting engine,” says Haribabu.

This ensures that cost reductions do not come at the expense of security, compliance, or operational risk, a critical requirement for industries like banking, healthcare, and government.

The Road Ahead: Cloud Cost Optimization as a Business Strategy

According to Haribabu that cost optimization is no longer an IT concern it’s a business imperative. As CFOs and CIOs increasingly demand granular visibility into cloud ROI, his solution positions optimization as a core enabler of digital profitability.

He is now working on the next phase autonomous workload optimization agents that will make intelligent decisions on behalf of systems based on evolving business objectives, compliance parameters, and real-time telemetry.

About Hari Babu Dama

Hari Babu Dama is a cloud database architect and optimization expert with over a decade of professional experience spanning Oracle, MySQL, PostgreSQL, MongoDB, and advanced multi-cloud ecosystems. Currently serving as an Application Architect IV at Randstad Digital LLC in Dallas, Texas, he has led high-impact database optimization projects for Fortune 500 clients including Bank of America and Wells Fargo.

His career began in academia, followed by progressive roles in enterprise IT where he specialized in complex database administration, performance tuning, disaster recovery, and cloud migration strategies. Hari Babu’s technical mastery extends across Oracle RAC, Exadata, GoldenGate, Azure, and AWS cloud environments.

He is also adept at automation frameworks using Ansible, Terraform, and Jenkins, integrating FinOps practices into DevOps pipelines for real-time cost and performance optimization. An alumnus of The University of Texas at Dallas with a Master’s in Business Analytics, Hari Babu bridges the gap between infrastructure and intelligence.

Article originally posted on mongodb google news. Visit mongodb google news

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


Podcast: Productivity Through Play: Why Messing Around Makes Better Software Engineers

MMS Founder
MMS Holly Cummins

Article originally posted on InfoQ. Visit InfoQ

Transcript

Shane Hastie: Good day folks. This is Shane Hastie for the InfoQ Engineering Culture podcast. Today I have the privilege of sitting down with Holly Cummins. Holly, welcome. Thanks for taking the time to talk to us today.

Holly Cummins: Oh, my pleasure.

Shane Hastie: InfoQ readers, QCon attendees will be familiar with you, but there’s probably four people in the audience who aren’t. Who’s Holly, tell us who you are and what got you where you are today.

Introductions [01:08]

Holly Cummins: So I’m Holly Cummins. I work for Red Hat and before that I was a long time IBMer. I think one of the joys of a career in technology is that I’ve done all sorts of things. So I’m currently part of the Quarkus team. I’m an engineer on the Quarkus team, but before that I worked as a consultant. So I was sort of helping organizations take advantage of the cloud, which was an incredible experience in terms of working with all sorts of different organizations and seeing some organizations that just were so incredibly effective. And then other organizations that maybe needed a little bit of help and needed a nudge in the right direction. And then I’ve been a performance engineer and I’ve worked on WebSphere and yes, all sorts.

Shane Hastie: Most recently at QCon London, you gave a couple of talks that I’d like to dig into. Let’s deal with the first one, productivity is messing around and having fun. Really?

The productivity paradox [02:08]

Holly Cummins: Yes. I mean that was a talk I did with Trisha G, and of course, we chose the title to be a bit provocative. It’s not intuitively obvious that that’s the case, but there is actually a lot of evidence that the way we think about productivity isn’t actually making us more productive. It’s making us less productive. And some of the things that we feel too guilty to do are actually the things that make us more effective. Because what we’re doing in technology, it’s not the same as working on a factory where we have an assembly line and we can just measure ourselves by how many things are coming off the end of the assembly line. We’re problem solving. And that’s fundamentally a creative activity. And one of the challenges we have because it’s a creative activity, because its knowledge work is it’s quite difficult to measure how effective we are.

We can’t just look and say how many widgets are rolling off the assembly line? And if we do it really wrong, we’ll try to and we’ll say, “Well, how many lines of code are running off the assembly line?” And that of course leads to some pretty terrible anti-patterns that I think probably all of us have seen where if you incentivize lines of code, you get lines of code. But lines of code are not an asset. Code is a liability. What you want is the minimum amount of code to solve the problem and the sorts of activities that allow us to solve the problem are not necessarily sitting down and typing as quickly as we can. There are things like thinking about how to solve the problem and where do we think best. We tend to not think best at a desk in front of a computer. We tend to think best doing things that are pretty apparently unproductive.

So a lot of us, for example, will solve problems in the shower. There’s something about being away in that environment where you can’t really do anything. You can’t bring your keyboard into the shower. And so then that just breathes a section of your mind to solve problems.

A lot of people find as well, things like loading the dishwasher or knitting or walking, that’s when you’re actually most productive. And it’s kind of good because I think there’s sort of a double win here because actually all of those things are sort of nice in their own right as well. So we’re in this really fortunate position where the things that make us more productive are actually the things that we kind of would want to do anyway. But the problem happens because they don’t look like productive activities. They look like we’re slacking off, they look like we’re just sort of loafing around. And so then it’s about how do we gather the narrative to give ourselves permission to do these things.

Shane Hastie: Yes, I don’t know whether my boss would be particularly happy with a timesheet entry that said showering, but you’re right, it is in this freer space, freer mental space. So we know this, but our organizations don’t seem to be getting this message. How do we communicate that?

The science behind creative downtime [05:15]

Holly Cummins: With all communication? Really it’s about figuring out the language of the person that you’re speaking to and communicating it to them in their terms. And in general, the business will want to see things in financial terms. So it’s quite useful to be able to actually quantify some of the benefits of this. But it’s also quite useful to be able to present the concrete evidence for it because there is quite a lot of evidence for it. So we can pull, for example, from psychological research because psychological research has actually been able to measure physiologically what’s going on when we have a shower, when we knit, there’s a part of the brain called the default mode network. So there’s a sub-discipline of psychology or what they do is they put people into an MRI machine and then they get them to do some task and they see which areas of the brain become more active.

And what they can see is that if they get people to do nothing at all, there’s one area of the brain or a network in the brain that becomes more active and it’s the doing nothing that enables activity in this area. And it’s called the default mode network. And so what’s the default in the mode network responsible for? It’s responsible for things like creativity, it’s responsible for things like problem solving.

Interestingly as well, it’s also responsible for something called mental time travel, which sounds amazing because it makes it sound like we’re going back and we’re being in medieval England, it’s nothing so exciting. It’s just when you replay what happened yesterday, that’s the default mode network. So we can sort of see at a physiological level what’s going on. We can also model it mathematically, which is kind of interesting. So again, if we need to communicate to someone who doesn’t want the sort of fluffy wuffy, “I feel nice when I do nothing”, who wants the, “Okay, what’s going on?”

Because if we look at queuing theory, that’s modeling how systems do work. And of course our organization is a system. We are a system. And what queuing theory tells us is that if you have a slightly random arrival rate of work, you need to have some period when the people doing the work are actually doing nothing at all.

Otherwise, what ends up happening is because there’s an asymmetry and if work is arriving faster than you can do it will just keep building up and up and up. And then the wait times in the system become well, infinite. You need to have some points in the system where the system appears to be doing nothing in order to keep the latency okay. And it’s exactly the same for us. So we need to have some periods in our day when we’re not doing anything or else any incoming task has potentially infinite latency, which isn’t what our management want.

Shane Hastie: But this is almost antithetical to what we’re seeing and hearing in the industry today. The push for more, do more with less, all of that.

“Do more with less” is counterproductive [08:22]

Holly Cummins: Yes, absolutely. And I think that’s why it’s really important to talk about this because the push to do more and more, more, it’s just wrong and it’s just counterproductive. I saw a social media post recently from someone and they were in the VC space and what they were doing was they had a group for their founders and I think it was WHOOP, maybe it wasn’t a tool that I’ve used, but they were tracking how much sleep their founders were getting. And they wanted their founders to be getting five hours of sleep a night because that shows that they really dedicated.

But again, this is something that there’s enormous amounts of evidence that if you go to work tired, that is equivalent to going to work drunk, you’re not achieving more, you are achieving less, you’re making more errors. And so if you said to people, “Would you like your employees to go to work drunk?” Everybody would be like, “No, that doesn’t seem like a good idea. I think I’d like to be a high quality organization”. But when you said, do you want your employees to go to work exhausted, people are like, “Oh yes, that shows were really..”. And it shows that productivity will be lower and the quality of life will be lower. So why would you want to have that double lose when you could be having the double win instead?

Shane Hastie: Which brings us to what is efficiency.

The efficiency trap [09:41]

Holly Cummins: Yes, and I think mean efficiency is interesting because we have sort of an intuitive definition of efficiency. When we talk about efficiency, what we tend to think of is, “Oh yes, efficiency means doing more”. But actually in a technical sense, efficiency is doing less. And if you want to be specific, efficiency is doing less in achieving the same results, which is obviously an important qualifier. But we tend to think of efficiency as ramming as much in as possible. And that’s not actually going to give us a good outcome. And again, we can see this for people and it’s sort of hard to accept for people because it seems a bit kind of fluffy.

But we see the exact same trade-offs and decisions being made in other fields where it’s more mechanical so it’s easier to accept. So for example, one lesson is that efficiency is always limited. So with an internal combustion engine, there is a top limit on its efficiency of about 36%. No matter how well engineered it is, it is never going to be more than that level of efficiency, which is irritating because you ideally would want more efficiency. Ideally we’d have like 120 efficiency. But then the other thing that’s really interesting is so we have this cap of 30%, but if you look at the efficiency of typical cars on the road, they’re not operating at that 36% level. They’re operating around more like 20%.

And is that because the engineers who designed those cars were sort of lazy and couldn’t be bothered? No, it was a deliberate decision to detune the engines because if the engines operate at that higher efficiency, they wear out faster. So you have this sort of short-term efficiency of how much fuel is going in and how much speed do I get out? And then you have this longer term efficiency of, “Does my engine actually last or do I have to replace my engine more often?” And the analogy with people is pretty obvious as well. If people work too much, it may seem like this short-term gain, but then you get burnout and that doesn’t help anybody.

Shane Hastie: There are times though when you do want the racing car and you’re prepared to throw the engine away.

When to push hard and the importance of recovery periods [11:50]

Holly Cummins: Yes, I think that’s right. And similarly, there’s times in the workplace where for all the sort of nice hand waving of, “Wouldn’t it be nice if we only worked at a sustainable pace? It only works seven and a half hours a day”. There’s times when you have a deadline, there’s times when you are racing against the competition when you need to get to market. So I think keeping that flexibility is important, but just making the conscious trade-off and saying this is going to be a period of high intensity, and then after that period of high intensity, I’m going to have a period of recovery. And again, actually that period of recovery can be very fruitful. So when I used to work in WebSphere, this was 15 years ago, the release cycles were quite different.

So you do a release every two years and it’s such a special occasion and if you miss that release train, you are not getting another chance for two years. So everybody is desperately trying to get into that release. And then of course then you sort of end up with this cycle where, I mean we’ve learned not to do that, right? Because everything gets shoved into the release. The quality is about the quality that you would expect if everything gets shoved in, so then you have to wait for the .1 or the .2 before you actually adopt it. And you think, “Um”. But leaving all that aside, what would happen is we would race to do this release working evenings and getting pizzas brought in and that kind of thing.

And then we would meet the deadline and then it would be like, “Okay”. And then what our management would say is, “Okay, we’re going to have a sprint where we are not going to be project managing this. You just do what you want to do”. Some of our best features came from those time off sprints because people could deal with the technical debt that had been bothering them or there was something that they just knew that would be really good, but they hadn’t managed to persuade project management to put it into the project plan. So they would just implement the feature and then it would go out and people would be like, “Ah, this is amazing”. So even that sort of apparent gift of time to work on what people wanted to work on ended up having quite big productivity benefits.

Shane Hastie: Shifting tac a tiny bit, before we started, I made the slightly tongue in cheek comments about we take the best engineer and we promote them and leave them to be the worst manager. How do we avoid that in our organizations?

Management vs. technical career tracks [14:13]

Holly Cummins: I think often that happens because in an organization, if the management track has higher advancement opportunities than the engineering track, people will be pragmatic and they will move across to that track. And so I think what we need to do is just make sure that we provide an opportunity for engineers who are really good engineers and really bad managers to stay in the engineering track, but be rewarded proportionally to the value that they’re bringing to the organization, which can be really quite significant.

Shane Hastie: And what about those that we want to go into that management track? What’s the advice for the engineer who is interested in going into management?

Holly Cummins: I mean, I have to answer this with caution because I have never been a manager and I have always been lucky enough to be in organization where there was a non-management track that went to quite senior levels. And so I’ve quite happily stayed on that track and never ventured across because I had a suspicion I wouldn’t be very good at it. But I think part of it then again, it’s about the opportunities for professional development, isn’t it? And so we see you have that with all of these things, you need to have that time to invest and then you get a higher reward.

And so that goes back to the efficiency conversation of maybe actually the way to get the highest value isn’t just in the short-term ring every bit of time out of people. We see it in the code base that maybe the way to get the biggest numbers of features isn’t to race through every feature and never put any time into technical debt. And it’s the same for people. Maybe the way to get the best person isn’t to sort of never give them any time for professional development, never give them any time for going off reading, learning. Because if we do that, we win in the short-term, but lose in the long term.

Shane Hastie: And advice for the person who wants to follow your path and stay in that technical… what should they, the more entry-level person who’s listening to this and looks at you and said, “I want to be like that”. What should they be focusing?

Staying technical: the curiosity factor [16:39]

Holly Cummins: I mean, I think so much of it is about the curiosity and keeping that curiosity always alive. I saw a quote recently and it said that creativity is when you do things from a place of curiosity rather than a place of fear. And I think this is such a creative industry, even if we don’t call it creative and it changes every single week. There’s new technologies, there’s new skills. And so we have to have that mindset where we’re embracing that newness and interested in what’s coming along. And also I think always looking to make things better as well.

I think sometimes, especially if we get a bit burned out or we get a bit ground down, you kind of look around and everything’s a bit terrible and there’s all this friction, you kind of go, “I can’t be bothered to change it”. And you have to have that interest in saying, “What’s not helping me? What’s not helping my colleagues? How can I fix it? Let’s make this happen”. And so sort of combining the curiosity and then let’s always be trying to make things better. Let’s be looking for the things that are bad. Let’s be continuously improving.

Shane Hastie: What are the trends that you’re seeing in our industry today that we should be thinking about and jumping onto?

The shift from author to editor role with AI tools and managing industry “fashion” [17:58]

Holly Cummins: I have mixed feelings. Having just said we should be curious and we should be embracing the new. I sometimes have mixed feelings about trends. A colleague of mine said yesterday that ours is a fashion industry and that’s completely true. So we do see some of these things come along where you have to pay attention to it in order to be in with the cool kids, but is it actually adding value? I’m not sure. But I think certainly one of the trends that we’re seeing at the moment is this move for all of us as engineers to be working higher up the stack and to be at the level where we’re managing code and managing code generation rather than writing it ourselves. And I think that’s an interesting transition that we need to figure out, “Well, what does my role look like now if I used to be a coder and now I’m a coder of coders, how does that feel? How do I manage that?”

Shane Hastie: I’ve described it and heard it described as the shift from author to editor.

Holly Cummins: Yes, I like that. And Annie Vella wrote this amazing blog where she was talking about exactly that. And in terms of our identity as software engineers, we came into it to do one thing and now we’re doing something else. And are we okay with that? Did I sign up to be an editor? Yes, it’s a period of change, I think for sure.

Shane Hastie: What are the underlying truths that aren’t changing?

Timeless truths in software engineering [19:37]

Holly Cummins: Oh, I like this. I think so much of it is not changing and we can get caught up in looking at these sizzles of change on the top and then miss all of the stuff that hasn’t changed. In our talk, Trisha and I referred a lot to Fred Brooks because if you read The Mythical Man-Month, so much stuff in there is exactly the same. And even stuff, maybe this has changed a little bit just in the last year, now that we’ve with that shift from author to editor, but even stuff like Fred Brooks wrote that you could expect to write 10 lines of code a day. And you look at that and you think, “That’s ridiculous. I can write 10 lines of code in five minutes. How?”

And then you sort of step back and you look at your output over the whole year, and then you look at how much of it is tests, infrastructure stuff, and then you look at how much of your time is going into meetings and conversations and design discussions. And then you divide that by 365 and you’re like, “Oh, actually that’s about 10 lines of code today”. So that side of it that there’s always overheads, there’s always, always overheads. How do we manage those overheads? Which overheads are the good overheads that we need? Because otherwise we just produce chaos. And which overheads are the overheads that actually we should be pushing back on and saying, “Why am I having to do this reporting? Why am I having to go to these meetings?”

And for both people and machines, that coordination just remains a huge challenge. So at the machine level, at the code level, it’s about the concurrency and how do we get the minimum amount of interlock that allows our code to function while still giving us good throughput and the number of cores is going up. That is a trend that we’re definitely seeing. So concurrency becomes more important. And then at the human level, it’s the exact same. How many meetings do I need to go to in order to ensure that my colleagues and I are pulling in the same direction without losing all of my time to meetings? So you can apply Amdahl’s law to code or you can apply Amdahl’s law to teams.

Shane Hastie: One of the things that we often hear is that the cognitive load on the engineer today is more than it used to be, and it’s getting more constantly. How do we manage that?

Managing cognitive load [21:58]

Holly Cummins: Yes, that one’s really hard. And I have such mixed feelings about it because when you look at a lot of the things that we’re being asked to do now, like shift left. On the one hand, shift left is an absolutely brilliant way of reducing these feedback cycles, improving the efficiency, getting away from that model where we sort of do the bad thing and then we throw it over the wall to the other team who says, you did the bad thing. And then we go back and we redesign it. But on the other hand, shift left for us is bringing it exactly as you say, all this cognitive load.

And there’s all these new skills that we need to know because now I need to be a deployment engineer because I need to understand where stuff is going. And if I want to have mechanical sympathy and write highly tuned code, then I need to have some hardware understanding as well. And I need to be a UX engineer and I need to have the front end skills and I need to do the security. Shift left should definitely apply to security. And then there’s so much. So I think really the only way that you can get rid of cognitive load is to, as more stuff comes in, you have to be getting rid of stuff.

And we’re seeing this a little bit, I think, because it used to be that as a software engineer you were working in such a low level, you were working in assembly, and now we can work in something that’s much more, even a modern high level language is so much more human-readable. So that piece of cognitive load has gone away. And a lot of our tools, we are sort of going up the stack. So in Java for example, we don’t need to be working with a low level concurrency constructs anymore. We can use structured concurrency or something like that. And so hopefully the ideal is for every piece of cognitive load that goes away, another comes in, but that they stay a bit in balance whether we’ve achieved it or not, I don’t know.

Shane Hastie: What are your hopes and dreams for our profession?

The joy of software engineering [23:56]

Holly Cummins: I think ours is such a fantastic profession. I feel so lucky to be a software engineer. So one thing that I do hope is that our industry carries on. There’s been a bit of concern now of will AI make us all disappear? And I don’t think that’s going to happen.

I think Jevons paradox means that our appetite for software is just going to continue growing. But I hope that at the moment, certainly in the right organization, this is such a joyful profession to be in. And we have such a nice combination of the problem solving and the creativity and the creation as well, which is sort of the same as creativity, but sort of not. And so I do hope that that will continue. And one thing I have really loved is the democratization of our profession. So we are seeing more and more people from different walks of life joining it. I really hope that continues because I think that’s so healthy for our industry.

Shane Hastie: Thanks very much. I’ve certainly enjoyed the conversation. If people want to continue the conversation, where can they find you?

Holly Cummins: So I’ve got a website at hollycummins.com, and there I sort of share talks. I blog, but not as often as I would like to, of course, I’ve got the backlog. But yes, I think that’s probably the best place. Or you can find me on Bluesky as well.

Shane Hastie: Wonderful. Thanks so much for taking the time to talk to us today.

Holly Cummins: Oh, thanks very much.

Mentioned:

About the Author

.
From this page you also have access to our recorded show notes. They all have clickable links that will take you directly to that part of the audio.

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


Multiverse names MongoDB’s Head of EMEA Donn D’Arcy as its new Chief Revenue Officer

MMS Founder
MMS RSS

Posted on mongodb google news. Visit mongodb google news

Training platform Multiverse has named Donn D’Arcy as is new Chief Revenue Officer, as the company reports accelerating growth.

D’Arcy joins Multiverse from database provider MongoDB, where he served as Head of EMEA, helping to scale the business to $700 million in Annual Recurring Revenue.

Multiverse says D’Arcy will help scale its goal to build the artificial intelligence (AI) adoption layer for enterprises and solve the workforce skills gap in AI. 

“Enterprise AI adoption won’t happen without fixing the skills gap. Multiverse is the critical partner for any company serious about making AI a reality, and its focus on developing people as the most crucial component of the tech stack is what really drew me to the organisation,” explains Donn D’Arcy, Chief Revenue Officer at Multiverse. “The talent density, and the pathway to hyper growth, means the next chapter here is tremendously exciting.”

Prior to joining MongoDB, D’Arcy spent more than 12 years at American company BMC Software, where he helped deliver UK revenue of $500 million, making it the company’s top-performing country worldwide. 

The appointment comes shortly after Jillian Gillespie joined as Chief Financial Officer, also from MongoDB and as Multiverse continues to drive business momentum. The company says its revenue “more than doubled” in the last two years and reports that more than 22,000 learners are now using its technology.

Euan Blair, Founder and CEO of Multiverse, comments: “Truly seizing the AI opportunity requires companies to build a bridge between tech and talent – both within Multiverse and for our customers. Bringing on a world-class leader like Donn, with his incredible track record at MongoDB, is a critical step in our goal to equip every business with the workforce of tomorrow.”

The news comes shortly after Multiverse announced plans to create 15,000 new AI apprenticeships over the next two years, aiming to address the UK’s skills gap.

Article originally posted on mongodb google news. Visit mongodb google news

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


Q1 Earnings Outperformers: DigitalOcean (NYSE:DOCN) And The Rest Of The Data Storage Stocks

MMS Founder
MMS RSS

Posted on mongodb google news. Visit mongodb google news

DOCN Cover Image

The end of an earnings season can be a great time to discover new stocks and assess how companies are handling the current business environment. Let’s take a look at how DigitalOcean (NYSE: DOCN) and the rest of the data storage stocks fared in Q1.

Data is the lifeblood of the internet and software in general, and the amount of data created is accelerating. As a result, the importance of storing the data in scalable and efficient formats continues to rise, especially as its diversity and associated use cases expand from analyzing simple, structured datasets to high-scale processing of unstructured data such as images, audio, and video.

The 5 data storage stocks we track reported a strong Q1. As a group, revenues beat analysts’ consensus estimates by 3% while next quarter’s revenue guidance was in line.

In light of this news, share prices of the companies have held steady as they are up 4% on average since the latest earnings results.

Weakest Q1: DigitalOcean (NYSE: DOCN)

Started by brothers Ben and Moisey Uretsky, DigitalOcean (NYSE: DOCN) provides a simple, low-cost platform that allows developers and small and medium-sized businesses to host applications and data in the cloud.

DigitalOcean reported revenues of $210.7 million, up 14.1% year on year. This print exceeded analysts’ expectations by 1%. Despite the top-line beat, it was still a mixed quarter for the company with a solid beat of analysts’ EBITDA estimates but EPS guidance for next quarter missing analysts’ expectations significantly.

“The momentum we generated in 2024 in both core cloud and AI continued into Q1, as we grew total revenue 14% year-over-year, our highest quarterly growth rate since Q3 2023, with AI ARR continuing to grow north of 160% year-over-year, and we delivered more than 50 new product features, over 5 times as many as we delivered in Q1 of last year.” said Paddy Srinivasan, CEO of DigitalOcean.

DigitalOcean Total Revenue

DigitalOcean delivered the weakest performance against analyst estimates and weakest full-year guidance update of the whole group. Unsurprisingly, the stock is down 15% since reporting and currently trades at $27.83.

Read our full report on DigitalOcean here, it’s free.

Best Q1: Commvault Systems (NASDAQ: CVLT)

Originally formed in 1988 as part of Bell Labs, Commvault (NASDAQ: CVLT) provides enterprise software used for data backup and recovery, cloud and infrastructure management, retention, and compliance.

Commvault Systems reported revenues of $275 million, up 23.2% year on year, outperforming analysts’ expectations by 4.8%. The business had a very strong quarter with a solid beat of analysts’ billings estimates and an impressive beat of analysts’ EBITDA estimates.

Commvault Systems Total Revenue

Commvault Systems delivered the biggest analyst estimates beat and highest full-year guidance raise among its peers. The market seems happy with the results as the stock is up 8.3% since reporting. It currently trades at $179.45.

Is now the time to buy Commvault Systems? Access our full analysis of the earnings results here, it’s free.

Formed in 2011 with the merger of Membase and CouchOne, Couchbase (NASDAQ: BASE) is a database-as-a-service platform that allows enterprises to store large volumes of semi-structured data.

Couchbase reported revenues of $56.52 million, up 10.1% year on year, exceeding analysts’ expectations by 1.7%. It was a satisfactory quarter as it also posted an impressive beat of analysts’ EBITDA estimates but a significant miss of analysts’ billings estimates.

Couchbase delivered the slowest revenue growth in the group. Interestingly, the stock is up 6.6% since the results and currently trades at $19.79.

Read our full analysis of Couchbase’s results here.

Founded in 2013 by three French engineers who spent decades working for Oracle, Snowflake (NYSE: SNOW) provides a data warehouse-as-a-service in the cloud that allows companies to store large amounts of data and analyze it in real time.

Snowflake reported revenues of $1.04 billion, up 25.7% year on year. This print beat analysts’ expectations by 3.4%. Aside from that, it was a satisfactory quarter as it also recorded an impressive beat of analysts’ EBITDA estimates but a miss of analysts’ billings estimates.

Snowflake delivered the fastest revenue growth among its peers. The company added 26 enterprise customers paying more than $1 million annually to reach a total of 606. The stock is up 18.1% since reporting and currently trades at $211.65.

Read our full, actionable report on Snowflake here, it’s free.

Started in 2007 by the team behind Google’s ad platform, DoubleClick, MongoDB offers database-as-a-service that helps companies store large volumes of semi-structured data.

MongoDB reported revenues of $549 million, up 21.9% year on year. This number surpassed analysts’ expectations by 4.1%. It was a very strong quarter as it also logged EPS guidance for next quarter exceeding analysts’ expectations and a solid beat of analysts’ EBITDA estimates.

The company added 110 enterprise customers paying more than $100,000 annually to reach a total of 2,506. The stock is up 2.1% since reporting and currently trades at $204.

Read our full, actionable report on MongoDB here, it’s free.

Market Update

Thanks to the Fed’s rate hikes in 2022 and 2023, inflation has been on a steady path downward, easing back toward that 2% sweet spot. Fortunately (miraculously to some), all this tightening didn’t send the economy tumbling into a recession, so here we are, cautiously celebrating a soft landing. The cherry on top? Recent rate cuts (half a point in September 2024, a quarter in November) have propped up markets, especially after Trump’s November win lit a fire under major indices and sent them to all-time highs. However, there’s still plenty to ponder — tariffs, corporate tax cuts, and what 2025 might hold for the economy.

Want to invest in winners with rock-solid fundamentals?
Check out our Top 5 Growth Stocks and add them to your watchlist. These companies are poised for growth regardless of the political or macroeconomic climate.

Article originally posted on mongodb google news. Visit mongodb google news

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


Offline caching with AWS Amplify, Tanstack, AppSync and MongoDB Atlas

MMS Founder
MMS RSS

Posted on mongodb google news. Visit mongodb google news

In this blog we demonstrate how to create an offline-first application with optimistic UI using AWS Amplify, AWS AppSync, and MongoDB Atlas. Developers design offline first applications to work without requiring an active internet connection. Optimistic UI then builds on top of the offline first approach by updating the UI with expected data changes, without depending on a response from the server. This approach typically utilizes a local cache strategy.

Applications that use offline first with optimistic UI provide a number of improvements for users. These include reducing the need to implement loading screens, better performance due to faster data access, reliability of data when an application is offline, and cost efficiency. While implementing offline capabilities manually can take sizable effort, you can use tools that simplify the process.

We provide a sample to-do application that renders results of MongoDB Atlas CRUD operations immediately on the UI before the request roundtrip has completed, improving the user experience. In other words, we implement optimistic UI that makes it easy to render loading and error states, while allowing developers to rollback changes on the UI when API calls are unsuccessful. The implementation leverages TanStack Query to handle the optimistic UI updates along with AWS Amplify. The diagram in Figure 1 illustrates the interaction between the UI and the backend.

TanStack Query is an asynchronous state management library for TypeScript/JavaScript, React, Solid, Vue, Svelte, and Angular. It simplifies fetching, caching, synchronizing, and updating server state in web applications. By leveraging TanStack Query’s caching mechanisms, the app ensures data availability even without an active network connection. AWS Amplify streamlines the development process, while AWS AppSync provides a robust GraphQL API layer, and MongoDB Atlas offers a scalable database solution. This integration showcases how TanStack Query’s offline caching can be effectively utilized within a full-stack application architecture.

Amplify Tanstack Optimistic UI - Interaction Diagram

Figure 1. Interaction Diagram

The sample application implements a classic to-do functionality and the exact app architecture is shown in Figure 2. The stack consists of:

  • MongoDB Atlas for database services.
  • AWS Amplify the full-stack application framework.
  • AWS AppSync for GraphQL API management.
  • AWS Lambda Resolver for serverless computing.
  • Amazon Cognito for user management and authentication.

Amplify Tanstack Optimistic UI - Architecture Diagram

Figure 2. Architecture

Deploy the Application

To deploy the app in your AWS account, follow the steps below. Once deployed you can create a user, authenticate yourself, and create to-do entries – see Figure 8.

Set up the MongoDB Atlas cluster

  1. Follow the link to the setup the MongoDB Atlas cluster, Database , User and Network access
  2. Set up the user
    1. Configure User

Clone the GitHub Repository

  1. Clone the sample application with the following command

git clone https://github.com/mongodb-partners/amplify-mongodb-tanstack-offline

Setup the AWS CLI credentials (optional if you need to debug your application locally)

  1. If you would like to test the application locally using a sandbox environment, you can setup temporary AWS credentials locally:
export AWS_ACCESS_KEY_ID=
export AWS_SECRET_ACCESS_KEY=
export AWS_SESSION_TOKEN=

Deploy the Todo Application in AWS Amplify

  1. Open the AWS Amplify console and Select the Github Option

Amplify Tanstack Optimistic UI - Select Github Option

Figure 3. Select Github option

2. Configure the GitHub Repository

Amplify Tanstack Optimistic UI - Configure Repository Permissions

Figure 4. Configure repository permissions

3. Select the GitHub Repository and click Next

Amplify Tanstack Optimistic UI - Select Repository and Branch
Figure 5. Select repository and branch

4. Set all other options to default and deploy

Amplify Tanstack Optimistic UI - Deploy Application

Figure 6. Deploy application

Configure the Environment Variables

Configure the Environment variables after the successful deployment

Amplify Tanstack Optimistic UI - Environment Variables
Figure 7. Configure environment variables

Open the application and test

Open the application through the URL provided and test the application.

Amplify Tanstack Optimistic UI - Sample Todo entries
Figure 8. Sample todo entries

MongoDB Atlas Output
Amplify Tanstack Optimistic UI - Data in Mongo
Figure 9. Data in Mongo

Review the Application

Now that the application is deployed, let’s discuss what happens under the hood and what was configured for us. We utilized Amplify’s git-based workflow to host our full-stack, serverless web application with continuous deployment. Amplify supports various frameworks, including server side rendered (SSR) frameworks like Next.js and Nuxt, single page application (SPA) frameworks like React and Angular, and static site generators (SSG) like Gatsby and Hugo. In this case, we deployed a SPA React based application. We can include feature branches, custom domains, pull request previews, end-to-end testing, and redirects/rewrites. Amplify Hosting provides a Git-based workflow enables atomic deployments ensuring that updates are only applied after the entire deployment is complete.

To deploy our application we used AWS Amplify Gen 2, which is a tool designed to simplify the development and deployment of full-stack applications using TypeScript. It leverages the AWS Cloud Development Kit (CDK) to manage cloud resources, ensuring scalability and ease of use.

Before we conclude, it is important to understand our application’s updates concurrency. We implemented a simple optimistic first-come first-served conflict resolution mechanism. The MongoDB Atlas cluster persists updates in the order it receives them. In case of conflicting updates, the latest arriving update will override previous updates. This mechanism works well in applications where update conflicts are rare. It is important to evaluate how this may or may not suit your production needs, requiring more sophisticated approaches. TanStack provides capabilities for more complex mechanisms to handle various connectivity scenarios. By default, TanStack Query provides an “online” network mode, where Queries and Mutations will not be triggered unless you have network connection. If a query runs because you are online, but you go offline while the fetch is still happening, TanStack Query will also pause the retry mechanism. Paused queries will then continue to run once you re-gain network connection. In order to optimistically update the UI with new or changed values, we can also update the local cache with what we expect the response to be. This is approach works well together with TanStack’s “online” network mode, where if the application has no network connectivity, the mutations will not fire, but will be added to the queue, but our local cache can be used to update the UI. Below is a key example of how our sample application optimistically updates the UI with the expected mutation.

const createMutation = useMutation({
    mutationFn: async (input: { content: string }) => {
    // Use the Amplify client to make the request to AppSync
      const { data } = await amplifyClient.mutations.addTodo(input);
      return data;
    },
    // When mutate is called:
    onMutate: async (newTodo) => {
      // Cancel any outgoing refetches
      // so they don't overwrite our optimistic update
      await tanstackClient.cancelQueries({ queryKey: ["listTodo"] });

      // Snapshot the previous value
      const previousTodoList = tanstackClient.getQueryData(["listTodo"]);

      // Optimistically update to the new value
      if (previousTodoList) {
        tanstackClient.setQueryData(["listTodo"], (old: Todo[]) => [
          ...old,
          newTodo,
        ]);
      }

      // Return a context object with the snapshotted value
      return { previousTodoList };
    },
    // If the mutation fails,
    // use the context returned from onMutate to rollback
    onError: (err, newTodo, context) => {
      console.error("Error saving record:", err, newTodo);
      if (context?.previousTodoList) {
        tanstackClient.setQueryData(["listTodo"], context.previousTodoList);
      }
    },
    // Always refetch after error or success:
    onSettled: () => {
      tanstackClient.invalidateQueries({ queryKey: ["listTodo"] });
    },
    onSuccess: () => {
      tanstackClient.invalidateQueries({ queryKey: ["listTodo"] });
    },
  });

We welcome any PRs implementing additional conflict resolution strategies.

Article originally posted on mongodb google news. Visit mongodb google news

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


Scylla Launches ScyllaDB X Cloud – I Programmer

MMS Founder
MMS RSS

Posted on nosqlgooglealerts. Visit nosqlgooglealerts

The developers of ScyllaDB have announced an updated version of the managed version of its database that is aimed at meeting workloads based on demand.

ScyllaDB is an open source NoSQL database that’s compatible with Apache Cassandra. The developers of Scylla describe it as a much faster drop-in replacement for Apache Cassandra. 

scylladb

Scylla Cloud was already available on AWS and Google cloud as a fully managed NoSQL DBaaS that could run data-intensive applications at scale across multiple availability zones and regions. Its benefits include the extreme elasticity with the ability to scale terabytes of data ingestion and transaction processing in minutes, alongside 100% API compatibility with Apache Cassandra CQL and DynamoDB and real-time streaming data processing via native Kafka and Spark plug-ins. 

This latest update, ScyllaDB X Cloud is the latest version, and the developers say it is a truly elastic database designed to support variable/unpredictable workloads. In practical terms, this release adds the ability to scale in and out almost instantly to match actual usage, hour by hour. In a blog post announcing the new version ScyllaDB’s Tzach Livyatan said:

“For example, you can scale all the way from 100K OPS to 2M OPS in just minutes, with consistent single-digit millisecond P99 latency. This means you don’t need to overprovision for the worst-case scenario or suffer latency hits while waiting for autoscaling to fully kick in. You can now safely run at 90% storage utilization, compared to the standard 70% utilization.”

The new mode uses a different cluster type, X Cloud cluster, which provides greater elasticity, higher storage utilization, and automatic scaling. X Cloud clusters are available from the ScyllaDB Cloud application and API on AWS and GCP, running on a ScyllaDB account or your company’s account with the Bring Your Own Account (BYOA) model.

The new clusters are based on ScyllaDB’s concept of tablets, smaller parts of tables. Data in Scylla tables are split into tablets, smaller logical pieces, which are dynamically balanced across the cluster using the Raft consensus protocol. Tablets are the smallest replication unit in ScyllaDB, and the developers say they provide dynamic, parallel, and near-instant scaling operations, and allow for autonomous and flexible data balancing. They also ensure that data is sharded and replicated evenly across the cluster. 

ScyllaDB X Cloud uses tablets to underpin elasticity. Scaling can be triggered automatically based on storage capacity, and as capacity expands and contracts, the database will automatically optimize both node count and utilization. Users don’t even have to choose node size; ScyllaDB X Cloud’s storage-utilization target manages that. 

The use of tablets also supports running at a maximum storage utilization of 90%, because tablets can move data to new nodes so much faster, meaning ScyllaDB X Cloud can defer scaling until the very last minute. 

ScyllaDB X Cloud is available now. 

scylladb

More Information

ScyllaDB Homepage

Related Articles

ScyllaDB Optimizes Mixed Workload Latency

Scylla Adds DynamoDB-Compatible API

ScyllaDB Launches DynamoDB Migration Tool

Scylla DB Adds Materialized Views

Scylla DB Adds HTAP

To be informed about new articles on I Programmer, sign up for our weekly newsletter, subscribe to the RSS feed and follow us on Twitter, Facebook or Linkedin.

Banner

espbook

Comments

or email your comment to: comments@i-programmer.info

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.


Presentation: Evaluating and Deploying State-of-the-Art Hardware to Meet the Challenges of Modern Workloads

MMS Founder
MMS Rebecca Weekly

Article originally posted on InfoQ. Visit InfoQ

Transcript

Weekly: I’m Rebecca Weekly. I run infrastructure at Geico. I’ll tell you a little bit about what that actually means. I’ll introduce the concept, and ultimately how we made our hardware choices for the efforts that we’re going through. I’m going to give the five-minute soundbite for what Geico is. Geico was founded in 1936. It’s been around a very long time. The sole purpose of what we live to do is to serve our customers, to make their lives better when things go bad. How do we do that? What does that look like? Just so you have a sense, we have over 30,000 employees. We have over 400 physical sites from a network perspective.

My team is responsible for connecting all of those different sites from a networking perspective. Then we have our primary regional offices. Those are 21 offices. Those are also connected. Then we have six on-prem data centers today. We have a pretty massive cloud footprint. My team owns the hybrid cloud across that experience, from the vended compute and storage down. I do not own any aspect of the Heroku-like platform stack. I really am at the vended compute and vended storage down across the clouds. That’s where I try to focus.

Geico – Infra Footprint

In order to tell you how I got to making hardware purchases, I want to take you back through how Geico ended up with an infrastructure footprint that we have today. In 2013, Geico, like many enterprises, made the decision to go all in on the cloud. We are going to do it. We’re going to exit our on-prem data centers and we’re going to go all in on the cloud. At that time, that’s when we had our six physical data center sites and all of our regional footprint, as I mentioned before. The reason why was maybe an interesting thought process. I’ll share it a little bit. I wasn’t there. What my understanding of the decision was, was a desire to move with more agility for their developers. They were feeling very constrained by their on-prem footprint in terms of their developer efficacy.

The thought was, if we go to the cloud, it has these fantastic tools and capabilities, we’re going to do better. This does not sound like a bad idea. The challenge was, they didn’t refactor their applications as they went to the cloud. They lifted and shifted the activities that were running on an SVC style storage SAN and on an on-prem footprint with old-style blade servers and separate L2, L3 network connectivity with subdomains all over, and a network segmentation strategy that really is something to behold. They moved that to the cloud, all in, 2014.

Fast forward to 2020, at that point, they were 80% cloud-based. Nearly 10 years into the journey, got almost all the way there. Prices had gone up for serving approximately the same load by 300%. Reliability dropped by two nines because of their surface area. I love the cloud. I worked at a cloud service provider. I am a huge fan of the cloud. Please take this with all due respect. It had nothing to do with a singular cloud. It had everything to do with how many clouds were selected. Every line of business chose their cloud du jour, generally associated with the anchor that they had. If they were using EDW, they’re going to end up in IBM.

If they were using Exadata, they’re going to end up in Oracle. This is how you end up with such a large proliferation and surface area, which when you add the complexity of wanting to have singular experiences for users, not shipping your org chart, creates a lot of latency, a lot of reliability challenges, egress fees, all sorts of fun, not-planned-for outcomes, which is how you go from a fairly flat compute load over a period of time, but get a 300% increase in cost. Not ideal, and a very large reliability challenge.

As I mentioned, not one cloud, not two clouds, but eight clouds. All the clouds. Including a primary cloud, absolutely. One that had more than half of the general spend, but every other cloud as well, and PaaS services and SaaS services, and many redundant services layered on top to try and give visibility or better composability to data. You just end up layering more things on top to try to solve the root problem, which is, you’ve increased your surface area without a strategy towards how you want to compose and understand and utilize your data to serve your customers. Should we start at the customer? That’s what our job is, is to serve them.

At that time, just to give you a sense of what that footprint became, just one of our clouds was over 200,000 cores and over 30,000 instances. Of that cloud, though, our utilization was on average 12%. In fact, I had scenarios where I had databases that had more utilization when doing nightly rebuilds than in the actual operation of those databases. Again, it’s a strategy that can be well done, but this is not the footprint of how to do it well.

What Changed?

To really dig into why, what changed, and how we got on this journey to look back and look at our infrastructure as a potential for core differentiation and optimization. One was the rising cloud costs. 2.5% increase in compute load over a 10-year period. 300% increase in cost. Also, when you look at the underlying features, there was one cloud in which we were spending $50 a month for our most popular instance type, was a VM instance type. That was actually running on an Ivy Bridge processor. Does anybody remember when Ivy Bridge was launched? 2012. I’m paying $50 a month for that instance. That is not even a supported processor. The kinds of fascinating choices that was the right choice when they moved, that was a current processor type.

Once people go to a cloud, they often don’t upgrade their instances, especially VMs, which will be more disruptive to the business, to the latest and greatest types. You end up with these massively slow, massively overprovisioned, potentially, to actually serve the business. Rising cloud costs, that’s how we got there. Number two, the premise of technology and developers being unlocked didn’t pan out. Why didn’t it pan out? Lack of visibility. Lack of consistency of data. Lack of actual appetite to refactor the applications. Lifting and shifting. We did things like take an ISV that was working on-prem, migrate it to the on-prem cloud, then that ISV later created a cloud offering, but we couldn’t take advantage of it because of where we were licensed and how we had built custom features on top of that. This is the story of every enterprise.

Every enterprise, whether you started with an ISV or an open-source stack, you start to build the features you need on top of it, and then it becomes very disruptive to change your business model to their managed service offering, their anything in that transition over. Unfortunately, it became even harder to develop for actually giving new services and features because we now had to add the elements of data composition, of SLOs across different clouds, increasing egress fees. The last thing was time to market. I just hit on it.

Ultimately, as a first-party tech group, our job is to serve a business. Their job is to serve our customers. If I can’t deliver them the features they need because the data they want to use or the model they’re looking at is in this cloud or this service model and my dataset is over here, I can’t say yes. The infrastructure was truly in the way. It was 80% of cloud infrastructure. That was where it was like, what do we have to do? How do we fix this to actually move forward?

The Cost Dilemma

I gave some of these numbers and I didn’t put a left axis on it because I’m not going to get into the fundamentals of the cost. You can look over on that far side, my right, your left, and see, at our peak, cost structure when we made the decision to look at this differently. We still had our on-prem legacy footprint. That’s the dark blue on top, and our cloud costs. You can do the relative more than 2x the cost on those two because, again, we hadn’t gotten rid of the legacy footprint, we couldn’t. We still had critical services that we had not been able to evacuate despite 10 years. Now, how am I changing that footprint? The green is my new data center that came up in July, and another new one that we’re starting to bring up on. You’re seeing part of the CapEx costs, not all the CapEx costs, in 2024. The final edge of the CapEx costs in 2026.

Then the end state over here is the new green for a proposed percentage. You all know Jevons paradox is such that as you increase the efficiency of the computation, people use it more. My assumption is the 2.5% growth rate, which is absolutely accounted for in this model, is not going to be the persistent growth rate. It’s just the only one that I can model. Every model is wrong, it’s just hopefully some are informative. That was the attempt here in terms of the modeling and analysis.

The green is the net-new purchase of CapEx. The blue is the cloud. I keep hearing people say we’re repatriating. We are looking at the right data to run on-prem and the right services to keep in the cloud. You can call that what you would like. I’m not a nation state, therefore I don’t understand why it’s called repatriation. We are just trying to be logical about the footprint to serve our customers and where we can do that for the best cost and actual service model. That is in perpetuity. We will use the cloud. We will use the cloud in all sorts of places. I’ll talk about that a little bit more in the next section.

As an insurer and as many regulated industries have, we have compliance and we have audit rules that mean we have to keep data for a very long time. We store a lot. How many of you are at a cloud service provider or a SaaS or higher-level service? If you are, you probably have something like an 80/20, 70/30 split for compute to storage.

Most people in financial services are going to look at something more like a 60/40 storage to compute, because we have to store data for 10 years for this state, 7 years for that state. We have to store all of our model parameters for choices around how we priced and rated the risk of that individual. We have to be able to attest to it at any time if there’s anything that happens in any of those states for anybody we actually insure. We don’t just have auto insurance. We are the third largest auto insurer. We also have life, motor, marine, seven different lines of business that are primaries, and underwriting. Lots of different lines of business that we have to store data associated with for a very long time. That looks like a lot of us. We have a lot of storage.

Storage is very expensive in the cloud. It is very expensive to recompose when you need to at different sites. That entire cycle is the biggest cost driver of why that cloud is so expensive. I want to use the cloud where the cloud is good, where it’s going to serve my end users with low latency in all the right ways.

If I have somebody who is driving their car as a rental car in Italy, I don’t build infrastructure anywhere outside of the U.S. I need to have the right kinds of clouds for content dissemination at every endpoint where they might use their insurance or have a claim. There’s always going to be cloud usage that makes perfect sense for my business. This is not the place where we want to be guardrailed only to that, because it’s a regulated low margin business. Reducing my cost to serve is absolutely critical for being able to actually deliver good results to my end users. This is the cost dilemma that I was facing, that our leadership team was facing, that as we looked at it, we said, this is what we think we can do given the current load, given what we know about our work. That’s how we got the go ahead.

Hybrid Cloud 101

I’m going to start with, how do we make a decision? How do you actually do this at your company? Then, go and delve into the hardware, which is my nerd love of life. First, alignment on the strategy and approach. I just gave you the pitch. Trust me, it was not a four-minute pitch when we started this process. It was a lot of meetings, a lot of discussions, a lot of modeling, seven-year P&L analyses, all sorts of tradeoffs and opportunities and questions about agility.

Ultimately, when we got aligned on the strategy and approach, then it’s making sure you have the right pieces in place to actually drive a cloud migration and solution. You got to hire the right people. You got to find the right solutions for upskilling your people that want to do new things with you. That was not a small effort. There’s lots of positions within Geico tech across the board. That has been, I think, an incredible opportunity for the people coming in to live a real use case of how and why we make these decisions. That’s been a lot of fun for me personally is to build a team of awesome people who want to drive this kind of effort.

Number three, identify your anchor tenants and your anchor spend. What do I mean by that? I’ve now used the term twice, and I’m going to use it in two different ways. I’m going to use it this way for this particular conversation. Your anchor services or anchor tenants are the parts of your cloud spend you don’t want to eliminate. These are likely PaaS services that are deeply ingrained into your business. It may be a business process flow that’s deeply ingrained, like billing, that has a lot of data that you don’t necessarily want to lose. Or it might be something like an innovative experience. Earlier sessions talked about generative AI and talked about hardware selection for AI.

There are so many interesting use cases for the core business of serving our customers in their claims, whether that’s fraud detection and analysis, whether that’s interactive experiences for chatbots, for service models, where we want to take advantage of the latest and greatest models and be able to do interesting things for our developers. Those are the kinds of use cases that are tying us to various clouds. Maybe CRM services. Every business is different, but whatever they have, you need to work with your partners. The infrastructure has to serve the business. We don’t get to tell them what to do. They have to tell us what they need to do. Then we look for the opportunities to optimize the cost to serve across that footprint.

Identifying those anchor services, and then the data. How is the data going to flow? What do you need the data for? Who are the services and users, and what are they going to need? How do we keep it compliant? How do we keep it secure across the footprint? Those are much more difficult conversations. Because everyone wants their data right where they are, but massive cost savings by putting it on-prem. What needs to actually be there? How do you create the right tiering strategy with your partners? Which you start to talk about different language terms than many businesses use. They don’t know SLOs. That’s not their life. They aren’t going to be able to tell me a service level objective to achieve an outcome. They will give you a feeling or a pain point of an experience that they’ve had.

Then, I have to go figure out how to characterize that data so that I have a target of what we can’t get worse than for their current experience or where they have pain today, and so where we need to turn it to, to actually improve the situation. Modeling your data, modeling and understanding what is needed where. Then, making sure you have real alignment with your partners on the data strategy for, might be sovereignty, I don’t personally have to deal with sovereignty, but certainly, for compliance and audit, is absolutely critical. It requires a lot of time. It has a lot of business stakeholders.

It is the most often overlooked strategy whether you’re going to the cloud or coming from the cloud. It will cost you if you don’t take the time to do that correctly. It will cost you in business outcomes. It will cost you in your customer service experience. It will certainly cost your bottom line. This is the step everyone forgets. Know what they don’t want to let go of, because you will not serve them well if you take it away. Know what they need to have access to, and make sure it is the gold. It is the thing that they have access to always.

Now you got that. You’ve got a map. You’ve got a map of your dependencies. You’ve got a map of your SLOs. You’ve got a map of where you’re trying to go. Now you need to look at your technology decisions. You get to choose your hardware, your locations for your data centers, your physical security strategies, all sorts of fun things. Then, you got to figure out what you’re going to expose to your users to actually do the drain, to actually move things from one location to another. You need to really understand what you’re building, why you’re building it, and then how you’re going to move people to actually create the right scenario to actually execute on this cost savings and vision. Then you create a roadmap, and you try and actually execute to the roadmap. That is the overview of what I’m about to talk about.

1. Start with Your Developers

Me, personally, I always want to start with my developers. I always want to start with my customer. For me, infrastructure, we’re the bottom. We’re the bottom of the totem pole. Everybody is above us. I need to know my data platform needs. I need to know all my different service layer needs on top of the platform, whether it’s your AI, ML. Then, ideally, you also need to turn that into an associate experience and a business outcome.

This is a generic stack with a bunch of different elements for the data and control plane. No way that I can actually move the ephemeral load or the storage services if I haven’t exposed a frontend to my developers, particularly my new developers, that is consistent across the clouds. Start with a hybrid cloud stack. What are you offering for new developers? Stand it up tomorrow. It’s the most important thing you’re going to do in terms of enabling you to change what’s happening below the surface. We start there. We start with our developers, what we need to expose. Those are good conversations. What do they need? If you have one team that needs one particular service, they should build it. If you have a team that’s going to actually build a service that’s going to be used by 4 or 5 or 7 or 12 different applications, that’s one that belongs in your primary platform as a service layer. Kafka, messaging, that’s a good example of one that you probably are going to want to build and have generically available across clouds. That’s the way.

2. Understand your Cloud Footprint

I will actually talk about turning the cloud footprint into a physical footprint and how to think about those problems. Our most primary cloud had over 100 different instance types. This is what happens when you lift and shift a bunch of pets to the cloud. You end up with a lot of different instance types because you tried to size it correctly to what you had on-prem. The good news about the cloud is that there’s always a standard memory to compute ratio. You’re going to get 4 gigs or 8 gigs or 16 gigs or 32 gigs per vCPU. That’s the plan. That’s what you do. You have a good provisioned capacity. Note I say provisioned. Utilization is a totally different game, and how you actually measure your utilization is where you’re going to get a lot of your savings. It’s important that you understand provisioned capacity versus utilization. What did we do? We took that big footprint of 103 whatever different instance types and we turned it into a set of 3 primary SKUs.

A general-purpose SKU, a big mem SKU, and an HPC style SKU for all the data analytics and ML that is happening on our company’s footprint. That was the primary. Then we had a bunch of more specialty variants. I don’t particularly want to keep a bunch of specialty variants for a long time. Again, not where infrastructure on-prem is going to be your best cost choice in savings. For now, there are certain workloads that really did need a JBOF storage, cold storage SKU. This is a big savings opportunity for us. There were definitely reasons why we got to that set of nine.

This one, we could have a longer debate about. Your provisioned capacity of network to your instance is very hard to extract. Each cloud is a little bit different in this domain. You can definitely see how much you’re spending on your subnets. You can see the general load across. You can see the various elements of your network topology designed in your cloud. Correlating beside the provisioned capacity, which usually your instance type will tell you what you have, but actually understanding how much of that network interface you’re using in terms of your actual gigabits is very hard. Different clouds have different choices. You can do a lot of things with exporters. If you’re using any kind of OpenTelemetry exporter and you have those options, you can try.

The hardware level metrics that we who build on-prem are used to, Perfmon, everything that you can pull out of PMU counters, everything you can pull out of your GPU directly if you’re actually in that space, you do not get that. That is probably the hardest area. Got a good sense of compute, something to think about in general from a compute perspective turning cloud to on-prem, is that for cloud, you are provisioned for failover. You have to double it. Also, if your utilization is, let’s say, 8%, 12%, 14%, you can hope your users will get towards 60% in your on-prem footprint as you move them towards containers. There’s hope and there’s necessity. Nothing moves overnight. You can do some layering to assume you’re going to get better utilization because you’ll have better scheduling, because you’ll have more control.

Ultimately, you still have to leave a buffer. I personally chose to buffer at the 40% range. Everyone has a different way they play the game. It’s all a conversation of how fast you can manage your supply chain to get more capacity if you take less buffer. Double it and assume you’re going to lose 40% for failover for HA, for all the ways in which we want to make sure we have a service level objective to our end users for inevitable failures that are happening on-prem.

3. Focus On your Route to the Cloud

Let’s talk about the network. Enterprises have fascinating franken-networks. I did not know this, necessarily. Maybe I knew this as an associate working at a company in my 20s. What happens? Why did we get here? How does this happen? What used to happen before COVID, before the last 15 years of remote work, is people had to go to the office to do their work. The network was the intranet, was the physical perimeter. This is what most of these businesses were built to assume. You had to bring your laptop to the office and plug in to be able to do your service. Then, during COVID, people had to go remote. Maybe that was the only first time it happened. Maybe it happened 10 years before that because they wanted to attract talent that could move and go to different locations, or they just didn’t want to keep investing in a physical edge footprint.

Whatever the reason, most of those bolted on a fun solution three hops down to try and do what we would normally do as a proxy application user interface to the cloud. Assume you have a large edge footprint and a set of branch offices on, let’s say, MPLS, which is rather expensive, it can double your cost per gig easily. Some places you’ll probably see it 5, 6, 10 times more expensive per gig. You’re probably underprovisioned in your capacity because you paid so much for this very low jitter connectivity, which someone sold you. That’s on a cloud. I’m not counting that in my eight clouds. I just want you to know that MPLS is done through a cloud, and you actually don’t know that it’s distinct two or three hops away from where you are. Lots of failure domains because you’re probably single sourced on that.

Then, it gets backhauled to some sort of a mesh ring. You’ve got some mesh that is supporting, whether it’s your own cloud, whether it’s their cloud, there’s some mesh that is supporting your connectivity. That goes into maybe your branch office, because, remember, all your network protocol and security is by being on-prem, so you’ve got to backhaul that traffic on-prem. Then that will go out to a CNF, some sort of a colocation facility, because that’s probably where you were able to get a private network connection, which if you are regulated, you probably wanted a private network connection to your cloud provider. That’s the third hop. Now that’s maybe the fourth hop, it depends.

Then you go to a proxy layer where you actually do RBAC, identity access-based control. Does this user, does this developer, does this application have the right to access this particular cloud on this particular IP? Yes, it does. Fantastic. You get to go to that cloud or you get to have your application go out to the internet. I talk to a lot of people in my job at different companies.

Most of us have some crazy franken-network like this. This is not easy to develop on. Think about the security model you have to enforce. Think about the latency. Think about the cost. It’s just insane. This is your route to the internet. This is going through probably an underprovisioned network. Now you have to think through, where do I break it? How do I change the developer compact so that this is their network interface? Wherever they are, any of these locations, it goes to a proxy, it goes out to the network. That’s it. That makes your whole life simpler. There’s a lot of legacy applications that don’t actually have that proxy frontend, so you have to build it. You have to interface to them. Then you manage it on the backend as you flatten out this network and do the right thing. It’s probably the hardest problem in most of the enterprises, just to give you a sense of that network insanity.

4. Simplify your Network and Invest in Security at all Layers

Again, all those boxes are different appliances. Because you have trusted, untrusted, semi-trusted zones, which many people believe is the right way to do PCI. Makes no sense. In the cloud, you have no actual physical isolation of your L2 and your L3, so if you promulgated this concept into your cloud, it’s all going on L3 anyway. You’re just doing security theater and causing a lot of overhead for yourself, and not actually doing proper security, which would be that anybody who’s going across any network domain is encrypted, TLS, gRPC. You’re doing the right calls at the right level and only decrypting on the right box that should have access to that data.

That is the proper security model, regardless of credit card information, personal information. This security theater is not ok. It’s not a proper model to do anything, and it’s causing a lot of overhead for no real advantage. Nice fully routed core network topology. Doesn’t have to be fully routed. You can actually look at your provisioning, your failure rates, your domains and come up with the right strategy here. That is not the right strategy. Maybe I’ll put one more point on it. Once you look at this franken-network you have and the security model you have, regardless of the provisioned capacity that somebody is experiencing in the cloud, it’s usually going to be the latency that has hit them long before the bandwidth. There is a correlation, loaded latency to your actual bandwidth.

Fundamentally, the problem to solve is the hops. Make the best choice from a network interface card and a backbone as possible. Interestingly enough, because the hyperscalers buy more at the 100-gig plus increments, generally, if I could have gotten away with 25 gig, if I could have gotten away with lower levels, go where the sweet spot of the market is. You’re not going to change your network design for at least five to seven years. It’s just not going to happen. Better to overprovision than underprovision. Go for the best sweet spot in the market. It’s not 400 gig, but 100 gig is a pretty good spot, 25 gig might be fine for your workloads and use cases.

5. Only Buy What You Need

Only buy what you need. I already gave you my rules around where you want to have doubling of capacity from your cloud footprint to your on-prem footprint, how you want to buffer and think through your capacity in those zones. When you’re actually looking at your hardware SKUs, very important to only buy what you need. I have a personal bias on this. I’m going to own it. A lot of people who sell to the enterprise sell a lot of services. Whether that’s a leasing model. Whether that’s call me if you need anything support. Whether that’s asset management and inventory. Whether that’s management tools to give you insights or DSIM tools to give you insights. These to me don’t add value. Why don’t they add value? Supply chain has been a real beast for the last four years.

If I’m locked into every part of my service flow, running on somebody’s DSIM that only supports their vendor or doing a management portal that only supports them as a vendor, I have lost my ability to take advantage of other vendors who might have supply. When I say only buy what you need, I mean buy the hardware. Run open source. It’s actually quite excellent. They’re probably using it if they’ve been using the cloud from a developer perspective, at least at the infrastructure layers. I’m not talking about PaaS layers. Truly at the infrastructure layers, they’re probably running Linux. They’re probably running more open-source choices.

If that’s the case, I personally looked at ODM hardware. I like ODM hardware. There’s ODM. ODM is a model, you can buy from somebody who’s a traditional OEM in an ODM style. That’s basically to be able to purchase the hardware that you want, to have visibility into your firmware stack, your BIOS, your maintenance, so that you actually can deploy and upgrade if you need to in post. Which is important to me, because right now I have a massive legacy footprint, but a bunch of developers building net-new stuff. What my memory ratios are right now may not be what they want to have in the next two years and three years, or storage, or fill in your note.

Doing this work, basically, and taking a model here of 1,000 cores, 1 terabyte of memory, yes, 1 petabyte of storage, so just normalizing out, we got about 50% or 60% less. That’s with all the bundling I mentioned. Double your capacity and buffer 40%, it’s still that much cheaper than those equivalent primary SKUs for vended capacity for compute and storage. That has nothing to do with PaaS, nothing to do with the awesome things in cloud. This is very specific to my workloads and my users. Your mileage may vary.

6. Drive your Roadmap

Finally, go take that puppy for a ride. You got to go do it. It’s great to have a model. It’s great to have a plan. You have to actually start that journey. We started our journey about June of 2023. The decisions were started and made in February, the analysis began, of 2023. We started making our first contact towards actually buying new hardware, actually looking at new data center location facilities in June of 2023. Issued out our RFPs, got our first NDAs signed, started our evaluations on those.

Actually, did our physical site inspections, made sure that we understood and knew what we wanted to contract for based on latency characteristics. By basically July of this year, we had our first data center site up and built on the new variety that is actually geo distributed. That was not actually taken into account in the six data centers we had previously. We’re a failover design. Then, had our first units delivered in September. Had everything up, running, debugged, and started to serve and vend capacity and compute through the end of this month for our first site.

Then, our second site coming up next year. Those didn’t happen overnight by any means. If I were to show you the hardware path of building out a new Kubernetes flow, of actually ramping and pushing up OpenStack for our fleet management on-prem, those were happening very similar timeframes, end of last year to build up the first, to really give a consistent hybrid cloud experience for our end users to onboard them there, running on top of public clouds, but getting away from the vendor locked SDKs into true open source. Then, giving us capabilities of later migrating so that you stop the bleed as you then prepare underneath the new hardware you want to have.

Lessons Learned

I have one more, things I wish I had whispered into my ear when we started this journey. There’s no amount of underestimation in terms of, you need the right team. To go from buyers to builders, you need the right folks who can do that. Doesn’t mean you can’t teach people. Doesn’t mean you can’t work together. You need the right senior technical talent and the right leaders in the right spots who’ve done this before. You need at least two years. You need a leadership team who understands it’s going to take two years. That they have to be wanting and willing. I’ve seen too many partners on this journey, or friends on this journey say, yes, it seems like a good analysis, but six months in, nine months in, new leader, we’re out. You’re not going to succeed in anything if you don’t have the willpower to stick with it. It’s going to be at least two years. Hardware and software don’t come together, or they shouldn’t. You need to really think through your software experience, your user experience, your hybrid cloud. You can do that now.

There are so many ways in which vendors get locked. You get locked into the services and the use cases of Amazon or Microsoft, or love them all. You can start to break that juggernaut immediately. You should for any reason. Whether you’re coming on-prem, whether you’re doing a hybrid cloud strategy, whether you want to find a different cloud for a specific use case. There’s a bunch of GPU focused clouds because it’s hard to get GPUs in the clouds. Whatever your reason is, understanding what is anchor and you’re going to keep it, and taking everything else out of the proprietary stacks gives you autonomy in making decisions as a business. If you care about margins whatsoever, it’s a good choice. Detailed requirements.

If there’s anything I found on this journey that I had to say to myself over and again is, do not let perfect be the enemy of dumb. It’s all going to change anyway. That’s the point. Take the best signal you can from the data you have, make the decision you have. Document your assumptions and your dataset and what might change it, and then go. Just go. Just try and create an environment for your team that it’s ok. That you’re going to screw up, it’s ok. Because there’s no way to get it right straight out of the gate.

The best thing you can do is to talk to your customers and make sure you really understand their requirements in their language. If you don’t have those conversations, you are definitely wrong. Maybe the other thing that is interesting is, open is not so open, depending on which layer of the stack you’re looking at, depending on even if you think you’re on the managed Kubernetes service that should in theory be the same as Kubernetes, no, it’s not. They’ve built all sorts of fun little things on the backend to help you with scaling.

Breaking it, even where you think you’ve chosen a more reasonable methodology, can be hard. I would be remiss if not saying, in this journey, there’s a lot of people who have helped us in the open-source community. That has been wonderful. Whether it’s CNCF and Linux Foundation, OpenStack, OpenBMC, Open Compute Project. This community of co-travelers is awesome. Very grateful for them. We’re members of most of these organizations.

Questions and Answers

Participant: The two years, the timeframe that you said, is it per data center?

Weekly: For me, that two-year roadmap is to go from six on-prem data centers to two data centers. Again, whether you do two or three is a choice for every company. You need two, because you want to have active-active. Unless you have a truly active-passive footprint, which maybe you do. Most companies want an active-active footprint, so you need two physical sites. If you have only two physical sites, you’re going to be writing your recovery log to the cloud. That is your passive site. That is your mirror. If you would rather do that on-prem, then you would want a third site. That’s a choice. It should come down to your appetite for spending in the cloud, where and why and how you want to think through your active-active and your recovery time. Cloud’s a great recovery place. It tends to have pretty good uptime when you have one of them. We’ve had some consternation given our journey and our experience in the cloud.

Again, I think that’s very much to the user, if you want to do three sites versus two. That’s the two years for the six to the two, or three. The longest end is usually the contracting on the frontend, doing the assessment, doing the site assessment, making sure they have the right capacity. Depending on what you’re purchasing from a data center colocation facility provider, I’m a huge fan of colocation facilities, if you are less than 25 megawatts, you don’t need to be building your own data center. Colos are great. They have 24-7 monitoring. They have cameras. They have biometrics. They are fantastic. They’re all carrier neutral at this stage. If you looked 10, 12 years ago, you might have been locked into a single service provider from a network perspective. All of them have multi-carrier at this stage. It’s a fantastic way.

Colos tend to have interesting pricing in terms of retail versus commercial versus high-end users, where you are actually having a colo built for you for your primary use. Most enterprises are going to be over retail, but way under commercial use. Pricing is different there than maybe other places. All the cost models I showed are very much in that over retail, if you’re under retail, the model does not hold. If you’re over retail size, they’re going to show pretty similar economics from a site perspective. Colo facility location buildout is usually 8 to 12 weeks. If you’re using a colo provider that doesn’t have a carrier you want to use, getting network connectivity to that site can be very time consuming. Outside of that, everything else is pretty easy to do.

See more presentations with transcripts

Subscribe for MMS Newsletter

By signing up, you will receive updates about our latest information.

  • This field is for validation purposes and should be left unchanged.