Hacker Newsnew | past | comments | ask | show | jobs | submit | apnew's commentslogin

hey Sumit!

Great seeing you here on and good luck!


Hey you, sorry I could not guess you from your username. Thanks a lot.


As someone struggling to quite literally fit into a similar role (have both words in official job title), I also feel most large company culture's are naturally poised to be at conflict with these roles!


This is one of those ghost jobs, the poster keeps posting a job but the job board has been broken all for forever. Someone replied to their prev posting and I even emailed them but received no response.

Ref: https://news.ycombinator.com/item?id=41730044


I've experienced the same with this job specifically. I wish forum mods would crack down on this as it's quite ridiculous


There should be a list of manufacturers who build durable consumer electronics.


Very very impressive.


Do you happen to have any handy guides/docs/references for absolute beginners to follow?


The absolute easiest way is https://github.com/Mozilla-Ocho/llamafile

Just download a single file and run it.


Ollama is not as powerful as llama.cpp or raw pytorch, but it is almost zero effort to get started.

brew install ollama; ollama serve; ollama pull llama3: 8b-v2.9-q5_K_M; ollama run llama3: 8b-v2.9-q5_K_M

https://ollama.com/library/dolphin-llama3:8b-v2.9-q5_K_M

(It may need to be Q4 or Q3 instead of Q5 depending on how the RAM shakes out. But the Q5_K_M quantization (k-quantization is the term) is generally the best balance of size vs performance vs intelligence if you can run it, followed by Q4_K_M. Running Q6, Q8, or fp16 is of course even better but you’re nowhere near fitting that on 8gb.)

https://old.reddit.com/r/LocalLLaMA/comments/1ba55rj/overvie...

Dolphin-llama3 is generally more compliant and I’d recommend that over just the base model. It's been fine-tuned to filter out the dumb "sorry I can't do that" battle, and it turns out this also increases the quality of the results (by limiting the space you're generating, you also limit the quality of the results).

https://erichartford.com/uncensored-models

https://arxiv.org/abs/2308.13449

Most of the time you will want to look for an "instruct" model, if it doesn't have the instruct suffix it'll normally be a "fill in the blank" model that finishes what it thinks is the pattern in the input, rather than generate a textual answer to a question. But ollama typically pulls the instruct models into their repos.

(sometimes you will see this even with instruct models, especially if they're misconfigured. When llama3 non-dolphin first came out I played with it and I'd get answers that looked like stackoverflow format or quora format responses with ""scores"" etc, either as the full output or mixed in. Presumably a misconfigured model, or they pulled in a non-instruct model, or something.)

Dolphin-mixtral:8x7b-v2.7 is where things get really interesting imo. I have 64gb and 32gb machines and so far the Q6 and q4-k_m are the best options for those machines. dolphin-llama3 is reasonable but dolphin-mixtral is a richer better response.

I’m told there’s better stuff available now, but not sure what a good choice would be for for 64gb and 32gb if not mixtral.

Also, just keep an eye on r/LocalLLaMA in general, that's where all the enthusiasts hang out.


Ollama is llamma.cpp plus docker If you can do without docker, it’s faster


Thanks for the book recommendations!


Same for Cannon pixma print plan. I have 2 brand new cartridges that I cant use because apparently they can be used only if I keep ~$6 monthly subscription which I absolutely dont need. HP and Cannon are on the black-list of brands.


When I was in HH, I loved the monthly pass offered by my employer (65€/month); the ability to travel longer distances over S-Bahn and taking +1 with you was killer/super-smart feature IMO. Miss those in states sorely.

I wish most, if not all, of the world adopts the german style public transport system.


This site adds a nice message about reader mode on iOS+Safari. Anyone knows how this can be achieved?


The text is in an element with CSS intended to make it “screen reader only” by visually hiding the element. Reader mode ignores those rules in the same way a screen reader would.

The unpleasant side-effect of this approach is that anyone using a screen reader for accessibility reasons is told to consider switching it off so they can look at the font the author picked.


Thank you for both insights!


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: