Hacker Newsnew | past | comments | ask | show | jobs | submit | milani's commentslogin

Often, when I use generative AI to produce videos, the results are close to what I envision but rarely capture my imagination exactly. Prompting the AI to fix specific details can be a cumbersome and time-consuming process. To address this, I'm developing solutions that make the creative workflow more intuitive. So far, I’ve built an app that allows users to provide visual clues as guides, along with a 3D environment where the camera can be freely manipulated within the generated scene.

The community is moving fast though. Now higgsfield allows using arrows and pointers to edit the video but so far, no one is doing a good camera control visually.


Not fulfilled yet? Maybe in the next tick.


Yeah, I'd heard they were still iterating on their routine.


Congrats on the launch!

I was scrolling and everything on the hardware looked great (although I was looking for how the tech is different from e-ink, not how better its properties are).

Then I reached "Introducing Sol:OS" and stopped right there. Didn't look further and closed the tab. I'm skeptical when it comes to custom OS. Not a lot of startups can handle a hardware+software company. Hardware is hard enough. Software is second class citizen and doesn't get regular updates after a few years.


It’s just an android rom tho?


Not comparing here, but I just remembered that the first paper on graphene was rejected (twice) and one of the reviewers said something like 'this paper does not represent a scientific advance.'


In my experience working with GPT4, if I give enough context on types, other functions definitions and the libraries I use, I get very accurate results. But it is a tedious task to copy paste from multiple places (type definitions, function definitions, packages, etc.).

In addition to the selected lines, does Continue support getting related definitions from the language server and inject them in the prompt? That would be huge.


This is very near on the roadmap, and we agree it will be awesome!

As of now, if there are collections of definitions that you frequently reference, you could save them in the system message, or write custom slash commands that let you prefix your prompt with these definitions.


> if I give enough context on types, other functions definitions and the libraries I use, I get very accurate results.

It's almost like .. coding it yourself!


After I copy every piece of code that is relevant, ask it to do something, then correct it's errors using human knowledge, then ask senior devs if it makes sense to revise my prompts and repeat, fix bugs from reviews from other humans, it's like this magical thing I get the mostly right answer!


For smaller pieces, you are right. But as a BE engineer, I was prototyping a concept and I didn't know much about Typescript+React+React-Router-Dom+React-Hooks-Form, etc etc. So I listed the libraries, a few lines of boilerplate and API definitions that RTK toolkit generated for me. Then asked GPT to generate the full page. It was much faster than I could "code it myself".

And that's why it is a "pain point". These all can be done automatically.


If you ever need to extend that full page to do something else or something additional you might find the time not so different after all. Because you still don't know much about Typescript+React+React-Router-Dom+React-Hooks-Form and now you have to read that stuff and try to make sense of it, even if just for passing it back into the bot.

Actually writing code is a minority of my time spent developing software, I don't need to trade that for need to spend time to know what and more importantly why something is where it is.

I guess it does work if you just want it to generate a webpage and you never want to then add any functionality. Or a glorified boilerplate generator for stuff not on the prod path.


> I don't need to trade that for need to spend time to know what and more importantly why something is where it is.

I had two options: 1. Learn everything first and then start prototyping. 2. Start prototyping and learn along the way. I chose the latter. But instead of searching stackoverflow and putting the pieces together, I used GPT and learned from the generated outputs in the context of my own problem.

It may not save a lot of time in larger projects because as you said, one should ultimately learn the "what and why", but it definitely provides a more pleasing experience. And I guess the time-saving part becomes more relevant with better tooling, like the suggestion I had.


Fair - I guess I'm just used to the old ways and hated things showing up without me specifically telling them to show up there first. But then, I still learn while prototyping, ChatGPT might be easier since it does give you a template to go off of.

But then, I don't think this really go beyond entry level prototyping, any complexity and chatGPT doesn't have enough tokens.


And driving to the store is basically just like walking there yourself!


I'd say it's more like driving vs assisted driving with shouting inputs (e.g. TURN LEFT! WATCH FOR THE PEDESTRIAN! CRUISE AT 70 MPH! SLOW DOWN THERE'S ANOTHER CAR IN FRONT! USE THE SECOND RAMP TO EXIT!). The only benefit is that you don't have to hold the wheel any more, the negative being that it still only works like 90% of the time, and you end up with laryngitis.


I have been experimenting a lot lately, and I would much rather copy paste high quality output(via providing context) than playing guessing games.

It's not like you have to be coding all the time.

Things will of course change as tools evolve.


Couldn't agree more—it's worth the extra effort to know exactly what enters the prompt. But control isn't mutually exclusive with removing the need to copy/paste. Continue lets you highlight code with perfect precision, and this is much lower effort.


I'd say it is a UX concern. It could show you the things it referenced in the prompt. So you'd only hit enter or customize it if needed :shrug:.


I also found this tedious and made a tiny vscode extension to make it less tedious

https://marketplace.visualstudio.com/items?itemName=TomJenni...


I don't think this is true. https://www.averagesalarysurvey.com/iran


These SEO-optimized salary websites are a joke, and usually never cite sources for their figures. An average annual salary of $17,600 makes Iran a relatively rich country...in reality, it's much lower than that. With crippling sanctions and extreme inflation, it should be somewhere between $100 and $500 monthly.


What is this? I am living in Iran and avg salary a month is 15 million tomans which is around 330 usd a month.


> lower cost as you scale

The cost aside, I'm wondering how fly or heroku support their customers when they grow to microservices ecosystems.

The problem shifts from deploying easily to deploying reliably meaning one release of a service should not break the other services. Other problems appear too, like service discovery, peer authentication, gateways, test and staging environments where there are downstream dependencies, etc.

Are customers supposed to leave when they grow to this level? Or are there solutions for these?


Look at this course from Pixar on storytelling https://www.khanacademy.org/computing/pixar/storytelling


Headsup: The link to Github Action does not work.

The difficult part for me is maintaining the docs over time. I can generate 80% of what theneo does automatically in swagger style. The remaining 20% (manual editing, what your platform offers) is the missing part with my current solution, yet I cannot see how theneo solves updating most of the doc and placing back the manual edits.


Hi milani, not sure I understood your question completely. For maintaining published documentation we offer: 1) User analytics & feedback 2) if you integrate with Github we allow auto-updates 3) Automatically send notifications about the update

Does this answer you question? Feel free to contact us through intercom if you want to go more into details as well.


What you mean when you say "I cannot see how theneo solves updating most of the doc and placing back the manual edits."?


What if you open an editor inside this terminal and collaborate in there?


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: