Hacker Newsnew | past | comments | ask | show | jobs | submit | tullie's commentslogin

Fair enough. We're releasing our first bring-your-own-cloud (BYOC) offering in April. We're working with a big e-commerce platform in Germany that has data sovereignty requirements so totally get the constraint and excited to offer something like this. We're starting with AWS then will do GCP at end of year. Full on-premise will still be awhile though to be honest.

For the cloud platform though (console.shaped.ai), i'd recommend just testing with some synthetic or deanonymized data or our demos and then if you're interested in BYOC reach out after April!


Makes sense. Implementation simplicity was part of the reason we didn't go for this. Currently the language is a transpiler, which maps SELECT and FROM into the retrieval stages, WHERE as post-filter stage, ORDER BY as score stage and REORDER BY to the final stage reranker. Because the text lookup is pushed through to the retriever we put it there. In the future we'll have more logic in the language compilation process which means that we can move things around a bit and under the hood they all get pushed to where they need to anyway.


I've been loving all these projects that are integrating LLMs/encoding directly into the language. There's so much power there.

Someone shared with me these the other day and we're inspired to add more remote LLM calls directly into ShapedQL now: https://github.com/asg017/sqlite-rembed https://github.com/asg017/sqlite-lembed


This line comes from a specific customer we migrated from Elastic Search, they had 3k lines of query logic, and it was completely unmaintainable. When they moved to Shaped we were able to describe all of their queries into a 30 line ShapedQL file. For them the reducing lines of code basically meant reducing tech-debt and ability to continue to improve their search because they could actually understand what was happening in a declarative way.

To put it in the perspective of LLMs, LLMs perform much better when you can paste the full context in a short context window. I've personally found it just doesn't miss things as much so the number of tokens does matter even if it's less important than for a human.

For the turbopuffer comment, just btw, we're not a vector store necessarily we're more like a vector store + feature store + machine learning inference service. So we do the encoding on our side, and bundle the model fine-tuning etc...


You can think of Shaped more like a vector store + feature store + ML inference combined into one service. This bundling is what makes it so easy to get state-of-the art real-time recommendations and search performance.

E.g imagine trying to build a feed with pgvector, you need to build all of the vector encoding logic for your catalog, then you need to build user embeddings, the models to represent that and then have a service that at query time encodes user embeddings from interactions does a lookup on pgvector and returns nearest neighbor items. Then you also need to think about fine-tuning reranking models, diversity algorithms and the cold-start problem of serving new items to users. Shaped and ShapedQL bundles all of that logic into a service that does it all as one in a low-latency and fault-tolerant way.


Thanks!


Subprocessors are here: https://docs.shaped.ai/docs/v2/support/security

Thanks for the feedback on the privacy policy, let me see if we can get that changed. For what it's worth we don't share personal information with anyone, this is likely just overly defensive legal writing on our part.


Makes sense! Agreed on the diversity for agents being a bit contrived here.

All of the retrievers do support pre-filtering, you just add the where clause within the retriever function. We're working on more query optimization to make this automatic also.


Would love to chat about it, and talk about dbt integration. There's a few use cases that have come up where this would be really helpful. I'll PM you.

1) So we do actually have a python and typescript API, it's just the console web experience is SQL only as it feels the best for that kind of experience. The most important thing though is that it's declarative. This helps keep things relatively simple despite all the configuration complexity, and is also the best for LLMs/agents as they can iterate on the syntax without doc context.

2) Yeah exactly, joins is something we can't do at the moment, and i'm not sure the exact solution their honestly. Under the hood most of Shaped's offline data is built around Clickhouse, and we do want to build a more standard SQL interface just so you can do ad-hoc, analytical queries. We're currently trying to work if we should integrate it more directly with ShapedQL or just keep it as a separate interface (e.g. a ShapedQL tab vs a Clickhouse SQL tab).

3) We didn't really want to create a new SQL dialect, or really a new database. The problem is none of the current databases are well suited for search and recommendations, where you need extremely low latency, scalable, fault-tolerance, but also the ability to query based on a user or session context. One of the big things here is that because Shaped stores the user interactions alongside the item catalog, we can encode real-time vectors based on those interactions all in an embedding query service. I don't think that's possible with any other database.

4) I haven't looked into mindsdb too much, but this is a good reminder for me to deep dive into it later today. From taking a quick pass on it, my guess is the biggest difference is that we're built specifically for real-time search, recommendations and RAG, and that means latency, and ability to integrate click-through-rate models and things becomes vital.

Thanks so much for the playground syntax, have some follow up questions but i'm going to pm you if that's okay. Agreed on the being able to see which columns exist.


Hey everyone,

I’m one of the founders of Shaped. We’ve spent the last few years building relevance infrastructure.

Over the time we've noticed a convergence happening: The stack for Search (usually Elastic/OpenSearch) and the stack for Ranking (Vectors + Feature Stores) are merging. Maintaining two stacks for the same math is inefficient.

We just launched Shaped 2.0 to collapse this stack. It treats Relevance like a database problem:

1. Ingest: Connect to Snowflake/Postgres/Kafka. 2. Index: We handle the embeddings (ModernBERT, etc.) and statistical feature engineering. 3. Query: We built a SQL-like language (ShapedQL) to retrieve, filter, and rerank candidates in <50ms.

Instead of writing Python glue code to merge BM25 and Vector results, you can do it in a single query:

SELECT title, description FROM semantic_search("$param.query"), keyword_search("$param.query")

ORDER BY -- Combine semantic relevance (ColBERT) with a personalized model

  colbert_v2(item, "$param.query") +
  click_through_rate_model(user, item)
We have a free tier ($300 credits/no-cc) if you want to try the SQL abstraction yourself.

Try it here: https://console.shaped.ai/register


Yeah exactly. I was really worried about reducing the serendipity that HN provides (as it's arguably why I've used it for so long as well) but the configurability allows it so that everyone can tweak their level of personalization to get their perfect goldilocks level.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: