

Im asking about this individuals experience as a ta, not for an opinion on llms.
Im asking about this individuals experience as a ta, not for an opinion on llms.
Oh interesting that they wouldnt need or want to hide that. When i use it i interpret every line of code and decide if its appropriate. If that would be too time consuming then i wouldnt use an llm. I would never deviate from the assignment criterion or the material covered by deferring to some obscure methodology used by an llm.
So i personally dont think its been bad for my education, but i did complete a lot of my education before llms were a thing.
Dont you guys test the students in ways to punish the laziness? I know you are just a ta, but do you think the class could be better about that? Some classes ive taken are terribly quality and all but encouraged laziness, and other classes were perfactly capable of cutting through the bullshit.
How do you know if it doesnt benefit a student? If their work is exceptional, do you assume they didnt use an LLM? Or do you not see any good code anymore?
Pavlov VR
So they are moving away from general models and specializing them to tasks as certain kind of ai agents
It will probably make queries with those agents defined in a narrow domain and those agents will probably be much less prone to error.
I think its a good next step. Expecting general intelligence to arise out of LLMs with larger training models is obviously a highly criticized idea on Lemmy, and this move supports the apparent limitations of this approach.
If you think about it, assigning special “thinking” steps for ai models makes less sense for a general model, and much more sense for well-defined scopes.
We will probably curate these scopes very thoroughly over time and people will start trusting the accuracy of their answer through more tailored design approaches.
When we have many many effective tailored agents for specialized tasks, we may be able to chain those agents together into compound agents that can reliably carry out many tasks like we expected from AI in the first place.
I think you hurt peoples feelings lmao.
The truth just isnt very catchy. Thanks for trying though. Im still on lemmy for people like you.
I think you are supposed to just do the homework
So an infant technology is showing a glimmer of maturation?
Cool ad?
Simulators and games with mods can push the cpu. But yeah. Mostly gpu limited.
It seriously cannot get any worse. Bring on users from any platform. Lemmy sucks and i thank the mods of smaller subs that post diligently about enthusiast topics to add content that isnt dystopian sad sack circlejerk
Take my job!!!
Im seeing a lot of reasons why you, or i, would not want such a service to exist.
What a person should or should not be doing is their business. Companies who can target vulnerable people would ideally be regulated.
Id much rather first go after payday advance companies with exorbant fees, or casinos, or high interest loans that individuals cant be expected to repay.
If someone came to a service provider and wanted it, and provided media to train on, and agreed to whatever costs are involved, isnt that enitrely their business?
How the fuck could this be illegal?
I dont understand how you dont question a jpeg. A child made this meme. It did not happen.
Now if it were a png, thats a good source. Theres a lot more transparency.
Oh okay. I was like… passthrough is a major part of the product lol.
Im amazed at the quality leaps for output for new versions of chatgpt.
I dont rely on it for project structure and the vast majority of logic but its so fucking cool that i can now give specific instructions and receive a mostly working script, give it the resulting errors, and after a couple prompts i have working code.
Not very surprising that people sticking their head in the sand have no idea what they’re talking about. Our methods of interaction and completeness of input are just as important as the capability of the model.
Is facial passthrough like face identification or just passthrough?
If you are making them aware they will fail by not reading the documentation, then its surprising they would continue to put that off. Using chatgpt is different than only being able to use chatgpt. Then again i was a kid once and kind of get it. Maybe banning it is the better option, as you say.
I thought it was scary enough when instructors would do “locked down” timed tests with short/essay answers. I cant imagine students thinking they’d be fine using chatgpt for stuff theyll need to applicably demonstrate.
I wonder if the drop out rate will increase for colleges due to stuff like this, or if students are majoring in more technical stuff more due to llm overconfidence.
Thanks for your responses!