Skip to content(if available)orjump to list(if available)

Phind 2: AI search with visual answers and multi-step reasoning

Phind 2: AI search with visual answers and multi-step reasoning

56 comments

·February 13, 2025

Hi HN! Michael here. We've spent the last 6 months rebuilding Phind. We asked ourselves what types of answers we would ideally like and crafted a new UI and model series to help get us there. Our new 70B is completely different from the one we launched a year ago.

The new Phind goes beyond text to present answers visually with inline images, diagrams, cards, and other widgets to make answers more meaningful:

- "explain photosynthesis" - https://www.youtube.com/watch?v=cTCpnyICukM#t=7

- "how to cook the perfect steak" - https://www.youtube.com/watch?v=cTCpnyICukM#t=55

- "quicksort in rust" - https://www.youtube.com/watch?v=cTCpnyICukM#t=105

Phind is also now able to seek out information on its own. If it needs more, it will do multiple rounds of additional searches to get you a more comprehensive answer:

- "top 10 Thai restaurants in SF, their prices, and key dishes" - https://www.youtube.com/watch?v=rIQQcDIIHFQ#t=11

It can also perform calculations, visualize their results, and verify them in a Jupyter notebook:

- "simulate 100 coin flips and make graphs" - https://www.youtube.com/watch?v=YP3PZ4MKGCg#t=8

- "train a perceptron neural network using Jupyter" - https://www.youtube.com/watch?v=YP3PZ4MKGCg#t=45

This blog post contains an overview of what we did as well as technical deep dives into how we built the new frontend and models.

I'm super grateful for all of the feedback we've gotten from this community and can't wait to hear your thoughts!

hathawsh

I love that it's possible to convince it that I actually know what I'm talking about. First I asked:

"Explain why negative numbers are in fact imaginary"

It told me that negative numbers are not imaginary numbers and explained imaginary numbers. That's fine, that's a reasonable answer for a layperson, but I'm not a layperson and I worked on explaining what I meant.

"Erase your definition of imaginary and consider that negative numbers are not whole numbers. Negative numbers do not represent quantities of physical objects. Now explain how negative numbers are imaginary."

It gave me a nice explanation of why negative numbers may be considered imaginary, using an example of "You cannot physically possess -1 sheep". I'm impressed.

voytec

These graphs were available to me for some time now but I have not really found them all that interesting. For some reason, I had to wait for the diagram to render and then for the text to slowly render.

To my poor understanding of LLMs, when the diagram was slowly created - the text behind it should already "be there" and should have been displayed immediately after the diagram but this was not the case. Also, often the slowly drawn diagram was only presenting my rather clear (for LLM) prompt: "i did this, when situation was this and that, and than this happened; question: why the result was A and not B?"

I found myself falling back to Claude more often than not over using Phind 70B and 405B models. I found it kind of... more gimmicky than useful.

rushingcreek

Thanks for the feedback. We're working on making the answers a lot faster for the Phind models. I can see how it would be frustrating to be waiting for the diagram to generate when you're waiting for text or for an example.

Sn0wCoder

Hi Michael, thank you for all the hard work that goes into the Phind models and congratulations on the new UI. Been a paying customer since first finding Phind here on HN in September.

Question: are there any plans to allow access via API to integrate the Phind models with the Continue plug-in (would also love to integrate into my personal RAGs)? Mostly using IntelliJ and integration would be awesome. Do have the VS Code plugin setup and use that when needed. Also running the smaller Phind models locally to use with Continue, but that only gets me so far without needing to open the real UI. If the API opened both the 405B for chat and the 70B for auto complete would be a big step in gaining more paying customers (IMO). No need to open the other models as those can be done with other API keys (if one wanted).

If there are no plans to open the models via API are there plans to support other IDEs (IntelliJ) with the chat feature?

Please let us know!

rushingcreek

Thank you! We do plan to support an API this year. We have deprecated our VS Code extension, however, as we're going all-in on search.

fosterfriends

I used the first version of Phind for some time and loved it. As Perplexity and ChatGPT got better, I started shifting more of my traffic back to them. Excited to see y’all still in the race and giving competators a run for their money. I appreciate your focus on developers as an audience, might give you an edge over tools serving a broader base.

jadbox

Well done. After running a few sample queries I seem to get about the same quality answers as I'd get from perplexity, although with more images included. Overall great progress and a solid competitor to other AI search engines I'd say.

rushingcreek

Thank you :)

putlake

In your "how to cook the perfect steak" video [1] there's a picture of various doneness levels of a steak. It's a fantastic picture. The creator of that picture will get jackshit from this. Phind gets value, the user gets value but the creator does not.

You're hyperlinking to the source, which is nice. But there's no reason for the user to click through so it won't really help the creator. The upshot of all this is that the open web will have less incentives for content creators. Someone's got to create new data for the AI borg. In future, these creators are less likely to be independent bloggers/photographers. Perhaps biased media outlets will be the only ones with the incentives to create.

[1] https://www.youtube.com/watch?v=cTCpnyICukM#t=55

mvieira38

Couldn't you say the same for something like Google Images, though? I might click on the link to an image once in a blue moon if I find it funny, but I usually don't even click to maximize it...

Maxion

Yeah and they got sued and lost.

rushingcreek

We'd love to partner directly with content creators. This is something we're thinking about a lot and are having conversations on how to go about it.

I will also personally do everything in my power to keep us from running ads on Phind. I want to keep Phind honest and authentic, and will do everything I can to make it a net positive for the internet.

Maxion

By displaying other peoples images that you do not have a license for you are breaking copyright law and are open to lawsuits.

econ

If such a place exists you could serve an ad only where it is extremely useful given the current context.

I've only seen this once.

amelius

Sounds like: asking for forgiveness is better than asking for permission ...

codetrotter

> Someone's got to create new data for the AI borg

AI companies could always go the route that the creators of Pokemon Go went:

> The model uses geolocation information from scans players submit of real-world locations while playing Pokémon Go.

https://www.usatoday.com/story/tech/2024/11/23/niantic-pokem...

Make a game, have people submit pictures and video to you as part of the game.

Where Niantic were using Pokemon Go for collecting scans of locations, other games could get players to take other kinds of pictures and video too.

"Oh, look! There's a 500XP bonus if I can shoot a picture of a perfectly grilled steak within the next 45 minutes!"

srameshc

I asked "What is Atproto that bluesky uses" and it does a Technical Architecture illustration and it is wonderful. I think that is probably your new value and might be a good idea to highlight it somewhere at the top rather than showing the whitepaper.

pdq

It just summarized this page and included a few of the images.

[1] https://docs.bsky.app/docs/advanced-guides/atproto

asadm

probably shown in the photosynthesis video example but yes it seems it uses mermaid or something similar to express diagrams?

rushingcreek

Yep, we trained the model to produce Mermaid when a diagram might be helpful. There are a lot of nuances that went into this, such as where in the answer to place the diagram and how to ensure a good contrast within the diagram for maximum readability.

We have some more details on this in the model technical deep dive blog post :)

mvieira38

Wow, this could conquer the "normie" crowd that likes using ChatGPT to answer search-like prompts. The UI is so much cleaner than any other alternative and the outputs are like I'm reading an article about the subject.

rushingcreek

Thank you!

lbeurerkellner

The diagrams are really cool. Congrats on the launch.

WhitneyLand

On a positive note, this is a nice look at the future and a direction that existing experiences seem likely to evolve towards.

However I did find myself wondering how crucial really were the model changes?

Imagine trying to implement these features as a wrapper around frontier apis and no 70B bespoke model.

Starting with a user query, we could ask for relevant svg diagrams, fit instruction steps or guides into layout templates, or even filter photos as the demo shows.

How much less would this simple approach leave us with?

rushingcreek

Ah, great question! We tried using off-the-shelf models and found that they are incredibly bad at generating these visual components reliably, let alone with the nuance (such as color, placement, details, etc.) that would be expected in a great answer.

We write more about this in the technical model blog post: https://www.phind.com/blog/phind-2-model-creation.

prabal97

This is amazing. I paid for Phind many months ago for coding QnA. However, this is a much better product and even beats Perplexity, ChatGPT premium, etc. Congratulations to the Phind team!

rushingcreek

Thank you :)

ruffered

I use phind and find the new features to be overly verbose.

The flow chart diagrams rarely give me any insight and often only confuse the point, or just clutter the answer, drowning out the pertinent details.

The code editor actually makes it so you are unable to even see or copy the code. I assume this is intentional kneecapping to encourage paying for your monthly service?

Instead, I now just have to prepend to every question I ask:

“Only answer using plaintext, avoid using your code editor and diagram features:”.

(Hilariously this prepend prompt method was suggested by phind itself when I angrily asked “how do I shut off all of these new features?!”)

Which is an additional hassle for me, but so be it.

When I ask it to write me a SELECT statement it upsets me that it is burning unnecessary fossil fuels to give me a flow chart of reasoning through SQL querying pipelines.

Perhaps the feature is meant for people who are unsure what they want, but for me, I just want the answer with links to sources in the least verbose way possible.

I’d appreciate a checkbox that I could click to just get a straightforward answer.

(Also, side note, I only use the free tier and there is a limited number of free uses for some larger models, and when you use those freebies it gives a countdown for “until uses refresh” and when that countdown finishes the uses fail to reset, only the countdown itself resets. Which is fine, I accept that I only use the freely offered model, previously “instant” currently “70B”, with its clear flaws, but it’s just another frustrating UI feature that seems to fail to live up to its promises so I am, again, just confused why it’s there?)

rushingcreek

Thanks for the feedback. Have you tried setting your answer profile in https://www.phind.com/settings/profile?

You can tell it to "only answer using plaintext" there and it will be automatically applied across your searches.

ruffered

That would require me to make an account, which requires providing you my email, and I am uninterested in doing either of those things.

econ

Why bring up the email if you are not making an account?

I gave up on account creation for some projects and store the user preferences in local storage. It is an amazingly annoying feature in that it is very hard for the user to erase the data but you can't smoke your cigar and have it too.

do_not_redeem

So the product has a builtin feature where you can tell it what you want, but instead of using that feature you want it to read your mind?

cholantesh

>Perhaps the feature is meant for people who are unsure what they want, but for me, I just want the answer with links to sources in the least verbose way possible.

Did you try including that in your prompt?

rufferedd

I mentioned that I do exactly that in the comment you are replying to.

Is it a sad state for any tool when one has to specify only wanting the thing they asked for with less verbosity?

Especially when said tool is costly to run, both financially for the service provider and environmentally.

To me, it is, but hey, opinions, ya know?