Hacker Newsnew | past | comments | ask | show | jobs | submit | ngalstyan4's commentslogin

Sounds cool!

Would be curious to know what the underlying aws ec2 instance is.

Is each DB on a dedicated instance?

If not, are there per-customer iops bounds?


We run on the same instance types the larger PlanetScale Metal sizes offer as whole instances. For Intel that's r6id, i4i, i7i, i3en, and i7ie. For ARM that's r8gd, i8g, and i8ge. (Right now, at least. AWS is always cookin' up new instance types.) Same story will soon be true for GCP.


there aren't per customer IOPs limits but the CPU will be the bottleneck.


They would need to handle all the translation changes as well, no?

<https://github.com/search?q=repo%3Apostgres%2Fpostgres+%22ma...>

I agree code change is simple, but I guess the task is complex for other reasons


If you just make a change of code, you don't need to handle translations at that time. That will get done by the various translation teams closer to the release. However you do need to make sure that the code is translatable (e.g. injecting pre-formulated english messages into a larger message is problematic).


Surprised to not see Ubicloud in there, which provides cloud services on top of various (including European) infra providers.

https://www.ubicloud.com/


They're under the CLOUD Act, i.e. quite toxic.


Basically everyone with even a CDN endpoint on the US is under Cloud Act. Hetzner, OVH, etc. Maybe only Scaleway that I couldn’t find any mentions of an US PoP.


US Cloud act says they need to provide US govt with access to your data even if the DC is outside the US


Not a European company are they though. Worst of both worlds as they come under US and EU regulation.


This is really cool, congrats on launch!

I am curious how you prevent private data from getting leaked to the auto-generated public docs. I imagine this problem does not exist in open source projects, but would become an issue if not everything discussed in company's private messenger should be used as context for generating docs.


Absolutely. There are steps in Promptless's agent flow that are designed to prevent this, but this is why users still review Promptless's suggestions to guides before committing/publishing them. I think people will still want to review Promptless's suggestions for a while, but the granularity of oversight will probably decrease as trust increases.


Author here.

> I don’t think “get moar ram” is a good response to that particular critique.

I do not think the blog post suggested "get more ram" as a response, but happy to clarify if you could share more details!

> Indexing in Postgres is legitimately painful

Lantern is here to make the process seamless and remove most of the pain for people building LLM/AI applications. Examples:

1. We build tools to remove the guesswork of HNSW index sizing. E.g. https://lantern.dev/blog/calculator

2. We analyze typical patterns people use when building LLM apps and suggest better practices. E.g. https://lantern.dev/blog/async-embedding-tables

3. We build alerts and triggers into our cloud database that automate the discovery of many issues via heuristics.


Author here. We will benchmark this thoroughly in the future for our vector indexes.

But at least anecdotally, it made a ton of difference.

We met <200ms latency budget with Ubicloud NVMes but had to wait seconds to get an answer from the same query with GCP persistent disks or local SSDs


We provide this functionality in Lantern cloud via our Lantern Extras extension: <https://github.com/lanterndata/lantern_extras>

You can generate CLIP embeddings locally on the DB server via:

  SELECT abstract,
       introduction,
       figure1,
       clip_text(abstract) AS abstract_ai,
       clip_text(introduction) AS introduction_ai,
       clip_image(figure1) AS figure1_ai
  INTO papers_augmented
  FROM papers;
Then you can search for embeddings via:

  SELECT abstract, introduction FROM papers_augmented ORDER BY clip_text(query) <=> abstract_ai LIMIT 10;
The approach significantly decreases search latency and results in cleaner code. As an added bonus, EXPLAIN ANALYZE can now tell percentage of time spent in embedding generation vs search.

The linked library enables embedding generation for a dozen open source models and proprietary APIs (list here: <https://lantern.dev/docs/develop/generate>, and adding new ones is really easy.


Lantern seems really cool! Interestingly we did try CLIP (openclip) image embeddings but the results were poor for 24px by 24px icons. Any ideas?

Charlie @ v0.app


I have tried CLIP on my personal photo album collection and it worked really well there - I could write detailed scene descriptions of past road trips, and the photos I had in mind would pop up. Probably the model is better for everyday photos than for icons


For similar isolation level anomalies in real world applications check out this SIGMOD '17 paper:

ACIDRain: Concurrency-Related Attacks on Database-Backed Web Applications: http://www.bailis.org/papers/acidrain-sigmod2017.pdf


Not sure what the approach of this library is, but can't you generate a nonce from a larger alphabet, hash the column values with the nonce `hash(nonce || column)`, and crypto-shred the nonce in the end.

Then, during hashing you just need a constant immutable state, which effectively expands the hash space, without incurring the mutable state overhead of replacement strings strategy.


this is cool!

Does this only collect logs from frontend?

Or it can also collect the backend and DB latency data related to a frontend interaction?


We collect logs across the stack. Here's some docs on our backend logging integrations (we also have connectors for major cloud providers): https://www.highlight.io/docs/getting-started#for-your-backe...


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: