Basically what the title says. I know online providers like GPTzero exist, but when dealing with sensitive documents, I would prefer to keep it in-house. A lot of people like to talk big about open source models for generating stuff, but the detection side is not as discussed I feel.
I wonder if this kind of local capability can be stitched into a browser plugin. Hell, doesn’t even need to be a locally hosted service on my home network. Local app on-machine should be fine. But being able to host it as a service to use from other machines would be interesting.
I’m currently not able to give it a proper search but the first glance results are either for people trying to evade these detectors or people trying to locally host language models.


There are no decent GPT-detection tools.
If there were they would be locally hosted language models, and you’d need a reasonable GPU.
I think I should have been more clear, this is exactly what I’m asking about. I’m somewhat surprised by the reaction this post got, this seems like a very normal thing to want to host.
Doesn’t help that some people here are replying as if I was asking to locally host the “trick” that is feeding a chatbot text and asking it whether it’s machine-generated. Ideally the software I think I’m looking for would be something that has a bank of LLM models and can kind of do some sort of statistical magic to see how likely a block of tokens is to be generated by them. Would probably need to have quantized models just to make it run at a reasonable speed. So it would, for example, feed the first x tokens in, take stock of how the probability table looks for the next token, compare it to the actual next token in the block, and so on.
Maybe this is already a thing and I just don’t know the jargon for it. I’m pretty sure I’m more informed about how these transformer algorithms work than the average user of them, but only just.