Showing posts with label computer science. Show all posts
Showing posts with label computer science. Show all posts

Thursday, June 26, 2025

My First Vibe Coding Hackathon

This past weekend my company did their first vibe coding hackathon. If you're not familiar with vibe coding, it's a new craze where you let AI Agents code your entire application and you just prompt the AI with "vibes." Pure vibe coding requires having zero coding knowledge, but developers have started finding ways to have the AI do most of the code writing while they review and analyze the output and make tweaks when needed.

Over the weekend, we were asked to use Replit and their AI Agent to create a working prototype of a useful application. The top 3 applications would be given money to the company store (which I personally enjoy since there are some quality items there) and we could present our applications to the company leadership. I threw my hat into the ring because why not! I hadn't vibe coded yet, but it would be good to know what's out there and the current state of AI.

This post won't go into the nuts and bolts of the app, but in about 7-8 hours I was able to build a working application that was 90% functional. It was a feature request tool that allowed your customers to request features they'd want to see in internal applications developed at their company. I took this idea actually from Compass Real Estate, where I used to work. They allowed their real estate agents to submit feature request for the Compass software tools. Much like Compass' tool, users of my application could:

  • Submit feature requests
  • Upvote other people's requests if they liked them (Reddit style)
  • Comment on feature requests
  • See new requests submitted in the past 48 hours
  • Review previous completed requests
Administrators had a special Admin Center page where they could:
  • Perform all the tasks in the above list
  • Consolidate duplicate feature requests
  • Add special "Admin comments"
  • Delete feature requests
  • Update the development status of feature requests
A screenshot of the main application page is below:



A screenshot of the admin center is below:


What I learned about Vibe Coding


AI Agents are not bad. They are not experts yet, but they do a pretty good job getting a basic application spun up really quick. This whole application took me two days, but really I spent 7-8 hours total doing everything. 

A lot of that 7-8 hours though was time spent waiting on the AI to finish analyzing, coding, reviewing, and processing all the changes I would request. While the AI is pretty good, it still can take 5-10 minutes to make a change, especially as the code base increases. 

Another good chunk of the time was taken up fixing bugs introduced by the AI when working on something completely unrelated. Several times the AI would mess something up in the application that was completely removed from what it was supposed to be working on. For example, when I had it work on the left-side navigation bar of the Admin Center, it also put the Update Status, Consolidate, and Delete button in one inline row and changed the button label for Update Status. Why??? I spent about 20 minutes getting the AI to fix the buttons without hurting anything else.

The AI Agents was not able to get the Search functionality working perfectly. I ended up flipping back and forth between two bugs in the search tool. When I fixed one issue, a second bug would appear. When I fixed that second bug, the first one came back. After a few rounds of that, I got the search working as best I could and left it at that. At the time of this writing, that is an active bug.

Another thing I learned was the AI is very verbose with writing code. This application has to be several thousand lines long. I feel like a human software engineer could write this in half the lines. But one nice thing about Replit's AI Agent at least is that it would give you an outline of what it was doing and give you a hyperlink to go in and check the specific file it is working on. I feel like if I was not rushing and trying to create as much as I can in one weekend, that is a very useful feature in the AI logic.

Overall Thoughts


Overall I felt more like a Product Owner this weekend and not a developer. I certainly didn't personally build this application, but I provided the ideas, creativity, prompts, and analysis that gave the application the look and feel. I wonder if that's what the developer profession will be one day - a world where we will all be product owners of some sort just requesting changes and troubleshooting manually when needed. Or maybe humanity will think "nah, this sucks. We want humans!" I hope it's the second option. 

Right now, the Agentic AI could get me nearly there. But it can't get this application to 100%. Advances are being made, and I'm sure in the future AI will be better and quicker. However for now, you still need some humans around. And I hope coding doesn't lose it's humanity any time soon.

Friday, June 6, 2025

PegaWorld 2025 (and my general thoughts on AI in 2025)

 


PegaWorld was a few days ago, and once again it was a huge blast. I'm fulfilling a promise to myself to actually write about it this year after neglecting to last year. I'll include a few things from 2024's edition though, as this year was sort of a sequel to the events from last year.

AI is here, and even low-code tools (which already do most of the code writing for you) are embracing it to the max. Last year, Pega announced several AI tools coming to their platform and this year followed up with the newest fad in AI - Agentic AI. For a platform that focuses on processes and workflows, I suppose this fits.

The conference itself

The conference itself, once you can get past some of the cheesy humor and jokes during the keynotes, is actually pretty fun. Alan Trefler's talk was a great introduction to the conference, and the point I took away from his talk was that you cannot stuff just any AI into a solution and think you will be okay - you need to use the right AI for the right problems. He compared it to attempting to use Chat GPT to help him with a chess puzzle. Chat GPT could not solve the problem, and instead gave Alan several incorrect/illegal moves in chess. As soon as Alan went to another AI that specialized in chess, it was solved in under two seconds. Thus begins a conversation about Pega having the right AI to solve workflow automation problems.

For the developer though, the keynotes are nice and some of the workshops and breakaway sessions can be really useful. But going to the innovation hub and seeing the new capabilities in action, talking with Pega engineers that built and specialize in that tool, and seeing how it can fit what we do on a day-to-day basis is the best part of the conference. I was able to meet some folks who are trying some new ideas in the Pega community, and I was able to learn about automated testing and network with people who have found creative ways to do it in Pega (a sticking point for our team right now).

I came back with several takeaways from the conference:

  1. Become a bigger part of the Pega community
  2. Pega Blueprint has changed quite a bit since being announced last year - I need to retrain myself on it and check out the latest updates.
  3. I want to learn more about automating testing within Pega via code scripts and accessing JSON files via their API.
  4. Pega Insights is a cool tool that we currently under utilize in our app at work - maybe we can use it?
  5. There will (hopefully) be a small Pega event in my city later this year, so I look forward to going to that.
  6. Pega recently released a whitepaper on AI in Operations, so I've downloaded that and plan to read it in the next week.

My AI thoughts in general

This conference both has me excited for the use of AI in application development and also terrified for the world we are creating. One of the Pega keynotes introduced their newest intern, Iris. Iris is a young girl who started working for Pega this year. She's fully remote, lives in Europe (Pega has a solid presence there, especially in the Netherlands), and has become a top contributor who has helped their employees with admin tasks - the type of stuff interns typically handle.

She's made mistakes on the job, as any person would. But to be frank, she's probably more of a professional and less green than I was when I was an intern. Iris loves her work, she loves where she lives (I believe it was an island with a northern lighthouse in the Netherlands?), and she enjoys working with her colleagues.

And as you've probably already guessed, Iris is also not real. She's completely made up - an artificially intelligent agent. 

On the one hand, this is cool. We've gotten to the point where computers can talk to us and be hyper-realistic. She has an employee picture, Pega employees refer to Iris with the she/her pronouns, and she has a personality that seems to make her human. 

But I cannot help but think we are starting to automate ourselves out of work. Right now, Agentic AI is being billed as the AI that will help you do the tasks you don't want to do. That's nice. But did Iris fill in a gap that would have been filled in the past by a real college intern? How about when Agentic AI gets to the point where it can do entry-level jobs - how long will it be when a CMO or VP of Finance decides that the entry level marketing or finance position can be handled by an Agentic AI instead of hiring a real human to do the same work? Because let's be real: computers have always been faster than humans at tasks that computers are capable of doing.

We're not far from this reality. Agentic AI, and AI in general, has made a lot of strides in the past few years since GPT 3 was announced to the world. And now, we have companies like Artisan in San Fransisco running marketing campaigns to stop hiring humans and instead hire their AI for massive cost savings. (Their marketing campaign is a sad one at that, purposefully created to make people angry and gather likes, clicks, impressions, etc. They should not have been rewarded for that, but alas here we are.)

All that to say, we may be coming to a point where we have to make some decisions with AI. How far are we willing to go to use it? Are we willing to let one AI handle the job of several people for the cost savings and speed? How many jobs are we willing to replace with AI if so? If all companies get on board with this, how quickly does it take us to automate 30% of the workforce out of a job? Will we make decisions today that harm the community at large over a long period of time in order to report good earnings and cost savings for the next quarterly investors call?

You can probably tell where I land on this. 

I was going to write about this in a separate post, but PegaWorld both last year and this year had me leaving with these apprehensions. I didn't come away from these keynotes fully excited for the future we are currently creating. I hope I'm wrong about it, I hope that companies still keep people employed and AI doesn't contribute to the death of several valuable industries and jobs. But after observing the past few years, I get ever so closer to feeling that we're all beginning to hand our jobs over to the control of just a few AI computers.

Wednesday, February 26, 2025

Creating an application using Vite

Up until recently, you could create a React application using Create-React-App. It was a lightweight way to create React applications quickly and was supported by the React documentation. Create-React-App has been discontinued - it will still work I suppose, but it is no longer the recommended way to create React applications.

I first came across Vite while going through freeCodeCamp's new curriculum. About a week later, while furthering my React studies on Scrimba, they also suggested using Vite as a worthwhile replacement. After building a few projects, I definitely agree.

Vite is a build tool for creating React applications. Of course, I could try to create my own React applications using no build tool, but that can take a long time and there are many files to configure. If someone else has done this work already, I'm going to take advantage of that! 

In order to use Vite, you need to install Node.js if you haven't already, and it needs to be a version higher than 18.0.0. Originally I had a higher version of Node installed (I think it was 22.0.0. or 23.0.0, I can't remember), and that caused some trouble with using Vite. Currently at the time I'm writing this, I'm using Node version 20.0.0 with zero issues.

Once Node is installed, do the following steps:

  1. Navigate in your terminal to the folder you want your project to be stored in. For me, I usually change my directory to Desktop and store my project there.
  2. Run npm create vite@latest 
  3. This will start the tool's prompts. It will ask you just a few questions:
    1. What is the project's name? 
    2. What library/framework are we using? 
    3. What language do you want to focus on? (For example, when I choose React for question #2, it will then ask if I want to focus on JavaScript, Typescript, or a few other options).
  4. Once these questions have been answered, Vite will build out the project. At the end, it will give you three commands to run in a row:
    1. cd <project-name>
    2. npm install
    3. npm run dev
  5. That last command will be one you probably use a lot, as it is the command that starts up the live server so you can see your work rendered in the browser. I personally find myself starting and stopping the server as needed, usually when I need to use my terminal for something.
And that's it! In the folder you selected, your new Vite project will be created. At this point, you can go into the project itself and clear out all the default stuff (styles, text, images, favicons) and begin creating your JSX files, stylesheets, and link them together.

Documentation link: https://vite.dev/guide/

Watching Claude play Pokemon

Currently Claude is playing Pokemon Red on Twitch. It was hard to tell what game it was playing, since the color palette of the game is very colorful with reds, greens, blues, and different shades of others (much unlike the real games being played in a blue or red hue). Now that I'm reading the Twitch "about" section, I see they are using Danny-E 33 + FroggestSpirit's full color patch which helps Claude see the screen better apparently.

This looks to be a passion project, likely a Claude super-fan or someone from Anthropic doing this on their own time. This is an experiment to see if Claude can actually beat the game on it's own with no outside help. It can learn the game, develop strategies, develop a battle theory when fighting other trainers, have goals for Pokemon levels, and learn the maps to move around.

I'm pretty fascinated right now while watching. Claude is moving very slow. Veeery slow. But I'm reading the left-side thought process that Claude has while seeing the game on the right, and I'm thinking to myself "This is exactly what my 9-year old brain was thinking when playing this game for the first time." I had Pokemon level goals, I had certain strategies to beat Brock's rock Pokemon at the very beginning (I had yellow and found out the hard way that Pikachu has zero use in that gym fight). It's pretty neat now to see an AI developing these same goals.

But as I mentioned above, it's very slow. I suspect in 5-10 years we'll have models that can play this game much quicker. But it soon gets a little boring watching the stream intensely, and the game music gets repetitive when listening to the same Viridian Forest theme for 30-40 minutes.

I've given it a follow and will be lurking to see how far we get. At the time of this writing, Claude has made it to Pewter City, and it has one Pokemon above level 10. That's expected for a player who has just entered Pewter City but it looks like it has several hours of training before it's ready to fight Brock.

You can follow along here: https://www.twitch.tv/claudeplayspokemon

Saturday, January 11, 2025

Reviewing HTML concepts and establishing a habit

 In my planner each week I write out three things automatically.

  1. Some sort of motivational theme or (soon) a bible verse as my "theme of the week"
  2. A task each day to exercise
  3. 30 minutes of freeCodeCamp
Exercising is still something that I need to improve on the habit front. Last week was a bit of a downer, and I think partially it's because I didn't drink enough water, exercise, and eat healthy (hello popcorn for lunch twice).

But with freeCodeCamp, I've been able to establish a nice rhythm. Either in the mornings or during my lunch break, I'll set aside 30-45 minutes to work on the new full-stack developer content they recently released (linked in my previous blog post). I decided to just do everything in the cert, even though I've previously finished their HTML/CSS content and most of their vanilla JavaScript content. It's good review, and I'm a completionist I suppose.

And I'm glad I've taken time to review the HTML content. First off, it's been way more in depth than their previous courses. There is a much higher focus on accessibility and semantic HTML. A few things which have stood out:
  • Presentational vs Semantic HTML: Presentational is older and includes elements that are no longer supported, and it's main focus is just on the look of the content on the screen. CSS has made most of these elements relatively pointless now. Semantic HTML involves using elements that are more descriptive in their purpose (as opposed to the older <div> element), and they are better for accessibility and website SEO.

  • HTML Forms have four states:
    • Default: which is editable
    • Focused: when you're in one of the form inputs specifically (i.e. clicking into a text field and the box highlighting a blue border)
    • Disabled
    • Read-Only
    • (The HTML Form states sort of mimic what I'm used to with Harness rules in Pega, so I found it interesting.)

  • HTML Entities: these provide a way to write out certain symbols that are usually code in HTML. For example, if I want to write out <div> in HTML text, without HTML entities I cannot accomplish this because the browser will think I'm trying to start a new <div> container. With HTML Entities, I can use a combination of letters and/or numbers to have the browser render symbols like < and >. These combinations are called references:
    • Named character references start with "&" and end with a ";"
      • &lt; is the < symbol
    • Decimal numeric references are similar but use numbers instead of letters
    • Hexadecimal numeric references are similar but use hexadecimal instead
I'm currently going through the lecture video series for accessibility, which is a deeper dive into screen readers and tools that come native with operating systems to help with viewing or listening to content. Additionally freeCodeCamp is reviewing various peripherals like larger-text keyboards and different input devices like trackpads and mouse-balls and joysticks. All must be taken into consideration for modern development.

That's all I have for this week. I'm also glad I could sit down for 20 minutes or so and type this out this week. I want to blog (ideally) every week, so this week was a success!

Saturday, November 16, 2024

Common String methods in Java

Almost all programming languages use Strings I'm sure. In Java, they are used a lot, and the language even made Strings it's own class. So, unlike other data types like `int` or `boolean`, Java's String data type is actually an object which is why we capitalize it when initiating a string.

Since String is an object, it also comes with it's own String methods. A full list of String methods can be found at this W3Schools link: https://www.w3schools.com/java/java_ref_string.asp

Below are some common ones:

  • length(): returns the length of a String as an int
  • concat(arg): returns a combination of String1 and the arg of the method - which could be another String or a mixture of blank spaces and a String.
    • String1.concat(" " + String2); as an example of multiple values in the argument
  • indexOf(arg): returns the index of the argument given
  • charAt(arg): returns a character at index arg
  • equals(arg): you cannot use the == operator to compare two strings in Java, so you need to use this method instead. This is case sensitive, but you can also use equalsIgnoreCase() as an alternative if you don't care about the case (or don't want to for the sake of your program's objective).
  • substring(arg, [arg2]): this returns a String that falls in line with the argument(s) you give the method signature. If you give just one argument, it will return a String that starts at that index all the way to the end of the String. If you declare two arguments, it will return a String that starts at the index of the first argument and will go up to but not including argument two.
  • toUpperCase()/toLowerCase(): these return a string in all caps or all lowercase. This is good for situations where you want to make sure case is ignored and you just want to compare the characters in the String.

Sunday, July 28, 2024

Writing my first SQL program

This was part of a Codecademy exercise in their intro to SQL course. Since I recently worked on my first story at work that required new SQL code to be written, I figured now is as good a time as any to start learning SQL.

 

Each step involves SQL "clauses" which are like keywords. They can also be referred to as "commands". The primary clauses here are:

  • CREATE TABLE
  • INSERT INTO
  • UPDATE
  • ALTER
  • DELETE FROM
  • SELECT

There are other helper commands being used like:

  • SET
  • WHERE
  • FROM
  • VALUES

These helper commands pair with the primary ones above to create SQL statements. Every statement ends in a ;. From experience, if you omit a semi-colon, the language will not run. This is much stricter than JavaScript.