X

Compare

Overview

Cursor AI vs Perplexity for math Comparison in different aspects of AI services with data mining from genuine user reviews & ratings, including: ALL,Interesting,Helpfulness,Correctness. AI store is a platform of genuine user reviews,rating and AI generated contents, covering a wide range of categories including AI Image Generators, AI Chatbot & Assistant, AI Productivity Tool, AI Video Generator, AI in Healthcare, AI in Education, AI in Lifestyle, AI in Finance, AI in Business, AI in Law, AI in Travel, AI in News, AI in Entertainment, AI for Kids, AI for Elderly, AI Search Engine, AI Quadruped Robot.

Rating Comparison

1 Reviews
0 Reviews

Reviews Comparison

Recommended Tags



  • codemonkey 2024-10-31 17:12
    Interesting:4,Helpfulness:4,Correctness:4

    Another more complicated task I tested on Cursor Code Editor is to write a python code to implement the Transformer function as in paper "Attention is all you need". The code I am writing include function name "def transformer(x):" and some of the beginning lines like "q=torch.linear(". And with the suggested code, I wrap up the transformer function in less than one minutes. So it can really help me increase the productivity of writing code. The only drawback is when I want to modify the code and use backspace to delete some lines, it seems like Cursor doesn't know what to do and give no hints for a while. Unless I start writing something else. And the newly suggested code will usually be quite different from the previously suggested ones. ``` ## Machine Learning ### Input import torch def transformer(x): ndim = 16 q=torch.linear(ndim, ndim)(x) k=torch.linear(ndim, ndim)(x) v=torch.linear(ndim, ndim)(x) attn = torch.matmul(q, k.transpose(-2, -1)) attn = attn / math.sqrt(ndim) attn = torch.softmax(attn, dim=-1) ### Cursor Suggested Code: out = torch.matmul(attn, v) out = torch.linear(ndim, ndim)(out) return out ```