please note at 13:46, mini got the answer right, as did llama3.3 at 14:11. i did point out earlier in the video, that position 2 is a correct answer. in the flow of the video i was focused on position 3, even though i had pointed out position 2 was also correct. you've all figured this for yourselves. i'm not editting this video as it doesn't change the narrative. apologies for missing this in the flow of the video
@Junon1529 күн бұрын
Saved me time and heartache figuring this out the hard way. I can ask for nothing more. Thanks!
@chrishayuk29 күн бұрын
super glad to hear it was useful
@patruff29 күн бұрын
No tool calling? Qwen will they be able to?
@RomuloMagalhaesAutoTOPO24 күн бұрын
Wow... very good and exploratory analysis. Thank you very much.👍 You reinforce the use specific model to specific problem, for example TOC to Agents... Thanks to remember this!
@3750gustavo29 күн бұрын
Nemotron got on first try when responding the '254*752-3+(2-7)+5': To evaluate this expression, we need to follow the order of operations, often remembered by the acronym PEMDAS (Parentheses, Exponents, Multiplication and Division from left to right, Addition and Subtraction from left to right). Here's how to break it down: Given expression: `254 * 752 - 3 + (2 - 7) + 5` 1. **Parentheses first**: Solve the expression inside the parentheses. - `(2 - 7) = -5` 2. **Update the expression**: Replace the solved parentheses with its result. - `254 * 752 - 3 - 5 + 5` 3. **Multiplication next** (since there are no exponents): - `254 * 752 = 191008` 4. **Update the expression again**: - `191008 - 3 - 5 + 5` 5. **Finally, handle Addition and Subtraction from left to right**: - `191008 - 3 = 191005` - `191005 - 5 = 191000` - `191000 + 5 = 191005` So, the final result of the expression `254 * 752 - 3 + (2 - 7) + 5` is **191005**.
@sgwong51324 күн бұрын
on PHI4, I got correct answer on first try if I give hint base on what I observe from the Nemotron output: calculate 254*752-3+(2-7)+5 To evaluate this expression, you need to follow the order of operations, often remembered by the acronym PEMDAS (Parentheses, Exponents, Multiplication and Division from left to right, Addition and Subtraction from left to right) Make sure you do the multiplication correctly, break it down and calculate multiplication step by step. Make sure you sum out multiple number correctly, break it down and calculate 2 number sum and repeat.
@JoshuaEworo18 күн бұрын
love the video! just found your channel too, definitely earned a sub! While it doesn’t support tool call via ollama it’s dead simple to use ollama “format: json” or “format: your schema” validation to do tool calling dead simply via python. I’d suggest exploring that further in a future video.
@pmarreck28 күн бұрын
FYI, on Mac at least, the BoltAI GUI app will connect to locally-running Ollama and LM Studio models served by those apps
@aaronabuusama15 күн бұрын
have you tried structured outputs with pydantic ai?
@jimlynch9390Ай бұрын
Are you sure selecting 2 is wrong?
@patruff29 күн бұрын
Google Fi, Phi, but when are they going to release Fo, and Fum? As an Englishman by blood please respond.
@chrishayuk29 күн бұрын
Hahaha, I should release the outtakes of the intro…
@stoicescucatalin807129 күн бұрын
What version of OpenWebUI are you using?
@artoke8429 күн бұрын
in openweb ui how did you setup a dark mode? it is so useful
@Maisonier29 күн бұрын
Great video. Liked and subscribed
@chrishayuk26 күн бұрын
Awesome thank you!
@DriftlessCryptoTooАй бұрын
Bravo!!! 🎉🎉🎉
@malllikvs29 күн бұрын
What is your hardware configuration ?
@alekseyburrovets47472 күн бұрын
in the tokenizer_config.json the fim tags are present. but in the ollama config ( application/vnd.ollama.image.template ) only chat assistant tags are present. why is that? which config to use? does the phi4 really have the fill-in-the-middle features?
@QuizmasterLaw29 күн бұрын
ollama run vanilj/Phi-4 Yes?
@chrishayuk29 күн бұрын
yep, ollama.com/vanilj/Phi-4
@QuizmasterLaw29 күн бұрын
@@chrishayuk thanks! liked and commented probably subscribed better check n b sure
@chrishayuk29 күн бұрын
thank you, glad you found the vid useful
@thegrumpydeveloper29 күн бұрын
I think Sam Altman said “no one wants last years model” or the second best model. Unless it’s really faster or better most will stick to one of the others. Good to see though.
@Tsardoz17 күн бұрын
You dont explain how to set up openwebui with phi4 unless it is already installed. I am having all sorts of trouble with that. Openwebui does not see phi4 at all even though it is running under ollama. PS. The issue I had was with docker under Linux not wanting to access localhost. I installed it using a venv and pip. Also ollama pull vanilj/Phi-4:Q8_0 ollama serve (probably not necessary - main thing was bloody docker)
@Corteum27 күн бұрын
If you're really good at using the model, could get a lot of work done with it, or does it have too many limitations and weaknesses?
@QuizmasterLaw29 күн бұрын
not yet available on huggingface library searching but if someone has a pull command from ollama please say
@thomecq29 күн бұрын
ollama run vanilj/Phi-4:Q8_0
@jacquesdupontd29 күн бұрын
uh ? ollama run vanilj/Phi-4:Q8_0
@henry-js27 күн бұрын
What are your pc specs??
@chrishayuk26 күн бұрын
macbook pro m3 max with 128GB of unified memory
@seniormcyt555227 күн бұрын
what's 254*752-3+(2-7)+5? first convert it to simpler parts then start solve it This model is very good at CoT, so if you make it think step by step, it gives you the right answer. For a better test of this model, you should make it break it down and think about it as much as possible. so the System Prompt should be : always break down any questions into simpler parts, then solve it step by step.
@chrishayuk26 күн бұрын
i shouldn't need to specify step by step for math problems... modern models know they need to do that for themselves.. so i purposely don't do it
@DrWaldonHendricks29 күн бұрын
I used a Gen10 A2 GPU, and it actually did a really good job on the latest model. It used about 10GB NVRAM at most compared with the o1, and it was not far of a really good model
@alelondon2329 күн бұрын
13:46 GPT4o mini got a RIGHT answer. position 2 is a winning move as much as 3 (you said it yourself earlier in the video and you mistakenly discarded the GPT4o mini answer as invalid. Again on 14:11 you wrongly stated Llama 3.3 gets it wrong when it gets it right! You had 1 job! Please fix the video or carry your reputation as a shill.
@chrishayuk29 күн бұрын
lol, you're correct, i got it wrong....
@chrishayuk29 күн бұрын
i've put a pinned comment pointing it out
27 күн бұрын
The Ollama modal is quantized to Q4_K_M, so you will lose much quality, especially context. It is not a fair comparison. Also, you may need some guidance to force thinking which o1 and Claude 3.5 are doing but Ollama doesn't support yet.
@chrishayuk26 күн бұрын
it's not a highly quanitized model, and i'm purposely focusing on areas where the quantization won't massively effect it. the model is not suddenly gonna gain more personality (this is a model size thing rather than quantizing), similar for the chains of thought, and similar for the code issues
25 күн бұрын
@@chrishayuk Q4 usually causes 5 to 10% drop in accuracy. Sometimes it can be up to 20%. Also, I said before, Ollama do not support forced thinking yet, so it is not possible to compare Phi-4 to Cladue 3.5 this way.
@d.d.z.29 күн бұрын
Nice video Chris
@chrishayuk29 күн бұрын
thaaaank you
@FalconStudioWin29 күн бұрын
The 14b parameter model may give worse answers in test time compute as smaller models generally does not reason better vs larger models