no. Grok lacks the memory. I don’t know why.
Source date (UTC): 2025-07-29 00:50:25 UTC
Original post: https://twitter.com/i/web/status/1949995894587150832
no. Grok lacks the memory. I don’t know why.
Source date (UTC): 2025-07-29 00:50:25 UTC
Original post: https://twitter.com/i/web/status/1949995894587150832
again, it needs training. I’m just kind of blown away at how good it is just using the books.
Source date (UTC): 2025-07-28 23:25:57 UTC
Original post: https://twitter.com/i/web/status/1949974640056566102
yes. that’s different from the provision of the answer. The AI was correct. But not correct enough for the precision you were looking for. Again, this is a matter of training and prompt configuration. The point is it still applies the method correctly and comes up with the right answer.
Source date (UTC): 2025-07-28 23:25:06 UTC
Original post: https://twitter.com/i/web/status/1949974425023000982
Mr. McMullan:
Correct me if I am wrong (because I’ve thought of this listening to Stefan lately), but Stefan works in philosophy, in particular, in choice and preference. While we work in truth, decidability, and law regardless of choice and preference. While we might argue in favor of minimum divergence from decidability under natural law, wouldn’t philosophical choice provide individuals with personal counsel where we would generally provide political counsel? (If you catch my meaning.) We do the via negativa (what not to do) leaving open the field of what one might choose to do … for the application of personal philosophy outside those negativa bounds.
I’ve been saying for a long time that we don’t do philosophy, we do science, but that the demarcation between philosophy and science wasn’t fully possible until our work. But it seems that for practical purposes we label ourselves as doing philosophy just for the sake of ease of communication.
I’ve been happy seeing Stefan back in the game because there is a need for it, and when people ask us we say that’s not our business. We don’t do preferable we do true and decidable. We only suggest you might prefer the decidable because it minimizes negative consequences. ;). That’s different from the choice of positivas which might seek to maximize the positive consequences. Yes? No? 😉
Source date (UTC): 2025-07-28 23:21:30 UTC
Original post: https://twitter.com/i/web/status/1949973518206030278
CurtGPT will give a tighter answer, and if then asked for the natural law case it will be very precise. I just tried and results as expected.
one of the current ‘limits’ to CurtGPT is that it doesn’t always use NLI methods if it can be answered conventionally. This is something that I need to update in the system prompts.
Source date (UTC): 2025-07-28 22:29:13 UTC
Original post: https://twitter.com/i/web/status/1949960361727131986
It does truth as well. the ‘big’ questions are just predominantly ethical and moral (political).
Source date (UTC): 2025-07-28 22:23:55 UTC
Original post: https://twitter.com/i/web/status/1949959027120648607
love you man. 😉
Source date (UTC): 2025-07-28 22:21:13 UTC
Original post: https://twitter.com/i/web/status/1949958350147293196
You only get to make a first impression once. I don’t expect to open it for public use until it’s been fully trained on at least Volume 2. Otherwise people will find edge cases and the dismiss the AI before it’s trained well enough to cover those edge cases. I just want to avoid negative impressions.
In a perfect world I would like it trained on volumes 1-4 before we went public. That’s only possible if we have funding to hire the team to perform the training.
I’ve been working on it today, but ChatGPT ‘got lost’ relatively quickly. And noticing it requires quite a bit of skill.
So I want the process down cold before we add trainers.
Source date (UTC): 2025-07-28 22:02:59 UTC
Original post: https://twitter.com/i/web/status/1949953758978330898
Maybe. I think my ‘acceptance’ of more of humanity has been most evident. I think the sense of hopelessness can easily set in. But humans occasionally surprise me. 😉
BTW: Hugs. 😉
Source date (UTC): 2025-07-28 21:44:58 UTC
Original post: https://twitter.com/i/web/status/1949949224533020795
I love my work, our work at the institute.
But oddly, the more we master our subject the more distant we become from the common man and woman, such that our very purpose – increasing trust, cooperation, justice, and quality of life – alienates us, even if only cognitively, from those who we struggle to serve. You don’t imagine that you’ll make that sacrifice when you start your journey. But you make it willingly once you begin serving.
Parenting a civilization is little different from parenting our children. The problem are just bigger, take longer, and are more challenging to solve. 😉
Source date (UTC): 2025-07-28 20:23:19 UTC
Original post: https://twitter.com/i/web/status/1949928677824946217