When the open source streaming service, Apache Kafka, was created in 2011 at LinkedIn, it was a different world.
The way they do that is by taking advantage of today’s cloud environment to separate compute from storage using an object storage service like Amazon S3.
“When you interact and store data in cloud object storage you get to sidestep all these networking fees that plague these big data systems when they get lifted, shifted into the cloud,” he said.
“Based on our experience building the kind of storage system on top of object storage we had built at DataDog, we felt like streaming systems should work the same way.
The founders brought some of the folks who helped build Husky to build the new system, and today they have 9 employees.
Continuous profiling reared its head in a 2010 Google research paper called: Google-Wide Profiling: A Continuous Profiling Infrastructure for Data Centers.
Polar Signals is the main developer behind Parca, a continuous profiling open source project which systematically tracks CPU and memory usage, creating profiles of this data to be queried over time.
“Our mission is to make the world’s datacenters ten times as efficient as they are today,” Polar Signals’ founder and CEO Frederic Branczyk told TechCrunch.
While cutting costs is one of the main benefits that Polar Signals promises, there are other benefits to the technology too — such as incident response efforts around a DDoS attack, for example, as Polar Signals can provide insights on the attack’s impact and identify which parts of a system are under stress.
At the time of writing, Polar Signals claims 11employees with experience at companies including AWS, Meta, Red Hat, and HashiCorp.
Jonathan Winer, the co-founder and co-CEO of SIP, said that the first three data centers designed using Verrus’ architecture will be located in Arizona, California and Massachusetts.
(Alongside the new business, SIP is also launching the Data Center Flexibility Initiative to bring stakeholders like energy companies, tech giants and regulators together in the meantime.)
Observing the strain that data centers in particular have on the electrical grid, SIP turned its attention to those data centers themselves.
Simply building more data centers, whether run by third-party data center operators or by the hyperscalers themselves, will not keep up with demand.
As SIP sees it, simply adding more data centers — which has been the approach up to now — is not a sustainable approach longer term.
Amazon’s cloud computing subsidiary AWS has revealed that it will allow customers to transfer their data out of its ecosystem with no so-called “egress fees” attached.
While AWS already allows customers to transfer up to 100GB of data per month off its servers for free, this won’t cover companies looking to “lift and shift” their entire data stores to another provider — and that is what is effectively changing for AWS customers as of today.
Companies that want to move their data off of AWS are requested to contact AWS, which will then apparently issue credits for the data being migrated.
It’s not clear what today’s news means in relation to the U.K.’s ongoing antitrust probe into cloud lock-in practices.
Another issue identified by the U.K.’s Competition and Markets Authority (CMA) was also looking at interoperability, concerning areas where cloud companies design their products to not play nicely with rival services.
Co-founders Ozgun Erdogan and Umur Cubukcu previously built Citus Data, where they also met Daniel Farina, who previously was instrumental in building Heroku PostgreSQL.
Citus Data wouldn’t have been possible without,” Cubukcu said.
“Fifteen years ago, all those service [that are on the Ubicloud roadmap] did not have open-source data plane components.
These days, for the core services, all of those now have good open-source alternatives, barring maybe one or two,” he said.
The company currently has 10 employees, split between San Francisco, Amsterdam (where the Citus Data team built a small engineering team during the team’s Microsoft days) and Istanbul.
A U.S. government watchdog stole more than one gigabyte of seemingly sensitive personal data from the cloud systems of the U.S. Department of the Interior.
The good news: The data was fake and part of a series of tests to check whether the Department’s cloud infrastructure was secure.
The experiment is detailed in a new report by the Department of the Interior’s Office of the Inspector General (OIG), published last week.
The tests were conducted between March 2022 and June 2023, the OIG wrote in the report.
The Department of the Interior manages the country’s federal land, national parks and a budget of billions of dollars, and hosts a significant amount of data in the cloud.
The launch partner for this is Google, which will use Stack Overflow’s data to enrich Gemini for Google Cloud and provide validated Stack Overflow answers in the Google Cloud console.
Google and Stack Overflow plan to preview these integrations at Google’s Cloud Next conference in April.
It’s no secret that content-driven services like Stack Overflow (but also Reddit, publishing houses etc.)
While Google and Stack Overflow aren’t discussing the financial terms of this partnership, it’s worth noting that this is not an exclusive partnership.
Google will also bring Stack Overflow right into the Google Cloud console and will allow developers to see answers and ask questions right from there.
When the large cloud providers have excess compute capacity, they tend to discount it through programs like AWS’s and Azure’s spot instances.
While in the recovery room, Surkov shared his idea for a company that would make this excess data center capacity available to developers.
Because of his experience with networking at Cisco and decentralized systems in crypto, Surkov had become interested in why there is so much friction in using decentralized compute.
In addition to these decentralized projects, NodeShift has also set up deals with independent data center operators.
Just on the data center side, the company says, it currently has access to about 400,000 CPUs and 28 million terabytes of storage.
Clumio, a data backup and recovery provider for companies using the public cloud, has raised $75 million in a Series D round of funding.
‘live”) data and backup data in separate secure domains, something that many backup solutions (including first-party from the cloud providers themselves) don’t do.
Indeed, Clumio stores its backup data in what Kumar calls a “hyper-optimized data plane” on AWS.
“We leverage our serverless data processing engine in AWS to perform all core backup operations using.
Prior to now, the company had raised around $186 million, the lion’s share arriving via a $135 million Series C round more than four years ago.
Interactive TC Early Stage sessions provide plenty of time to engage and ask questions.
She has 19 years of investing and operational experience in high-growth companies in IT software, cloud and enterprise SaaS.
TechCrunch Early Stage 2024 takes place on April 25, 2024, in Boston.
Buy a TC Early Stage pass now, save up to $300* — and join us in Boston!
Is your company interested in sponsoring or exhibiting at TC Early Stage 2024?