News

The New Stack
thenewstack.io > llm-database-context-mcp

How your LLM is silently hallucinating company revenue

2+ hour, 3+ min ago  (531+ words) LLMs are accelerating work across engineering disciplines, from generating React components and building backend APIs to noodling with SQL. But we all know LLMs make mistakes, and the nature of those mistakes varies dramatically across domains. Using LLMs with databases…...

The New Stack
thenewstack.io > multimodal-rag-hybrid-search

Your RAG System is probably image-blind, but it doesn't have to be

1+ week, 3+ hour ago  (187+ words) To unlock the true value of these assets, businesses must move beyond text-only RAG. The next frontier is multimodal RAG with hybrid search. Traditional RAG systems are often "image blind." They can process the text in a PDF but ignore…...

The New Stack
thenewstack.io > pg_lake-comes-to-snowflake-postgres-what-it-means-for-open-standards

pg_lake comes to Snowflake Postgres: What it means for open standards

1+ week, 6+ day ago  (276+ words) Snowflake Postgres unifies analytical and transactional workloads by allowing users to access PostgreSQL directly within Snowflake. Consequently, they can read and write to open table formats via SQL; pg_lake abstracts most of the underlying complexity required to manipulate data of different…...

The New Stack
thenewstack.io > meet-gravitino-a-geo-distributed-federated-metadata-lake

Meet Gravitino, a geo-distributed, federated metadata lake

3+ week, 3+ hour ago  (523+ words) Metadata has been a second-class citizen, according to Junping (JP) Du, founder and CEO of Datastrato, a data and AI infrastructure company. AI is changing how data " and metadata " is consumed, understood, and governed, so Datastrato created Apache Gravitino, an…...

The New Stack
thenewstack.io > prompting-vs-rag-vs-fine-tuning-why-its-not-a-ladder

Prompting vs. RAG vs. fine-tuning: Why it’s not a ladder

3+ week, 5+ hour ago  (1641+ words) Teams usually assume there's a straightforward progression from prompt engineering through retrieval-augmented generation (RAG) to fine-tuning (the last rung on the ladder) when customizing large language models (LLMs). This is an easy-to-understand, frequently repeated narrative that is true for some…...

The New Stack
thenewstack.io > art-of-visual-inspection-spot-the-hidden-story-in-your-charts

The art of visual inspection: Spot the hidden story in your charts

3+ week, 4+ day ago  (1110+ words) You have a nice set of data, but how do you convey its meaning to an audience? In the latest of a series of free technology talks, the Association for Computing Machinery presented noted researcher and author Angelica Lo Duca…...

The New Stack
thenewstack.io > llms-create-a-new-blind-spot-in-observability

LLMs create a new blind spot in observability

3+ week, 5+ day ago  (592+ words) Teams that have run Software as a Service (SaaS) products know the routine: An alert goes off, and everyone scrambles to the observability stack first. Metrics, logs, and traces have been the signals that help engineers figure out what broke,…...

The New Stack
thenewstack.io > how-precog-adds-business-context-to-make-enterprise-data-ai-ready

How Precog adds business context to make enterprise data AI-ready

4+ week, 9+ hour ago  (510+ words) Extracting data from enterprise tools like Salesforce, SAP Ariba, or NetSuite is relatively easy. Making that data usable for AI models to reason over is much harder. Just having massive amounts of tables and columns or giant multidimensional JSON files…...

The New Stack
thenewstack.io > ai-sre-roundtable-the-creation-of-a-new-category

AI SRE roundtable: The creation of a new category

4+ week, 1+ day ago  (401+ words) It's no surprise following the explosion of AI-driven code generation tools, that we are now seeing a strong AI-driven response from operations and security. Almost by definition, a core function of SREs and security professionals alike is keeping up with…...

The New Stack
thenewstack.io > why-d-matrix-bets-on-in-memory-compute-to-break-the-ai-inference-bottleneck

Why d-Matrix bets on in-memory compute to break the AI inference bottleneck

4+ week, 1+ day ago  (975+ words) AI inference is poised to become the single largest compute workload. For some time now, after all, AI service providers like OpenAI have claimed that they are mostly constrained by compute availability. Even as these companies build massive, power-hungry data…...