For most of college, I’ve kept it simple: I’d create a directory in my home folder for each project, then eventually move older or inactive ones into ~/programming/. When I change devices or hit file size limits, I’ll compress and send things to my NAS.

This setup has worked pretty well so far. But now that I’m graduating and my projects keep stacking up, I’m starting to wonder if there’s a more efficient system out there.

Curious—how do you all organize and store your projects? Any tips or methodologies that have made your lives easier over time?

The only person I’ve talked to about this is my mentor who’s been programming since the 60s (started on the IBM 1620 and Bendix G15) and he just mostly keeps projects in directories in his home directory and uses his godly regular expressions skills to find things that way. Makes me wonder if I’m overthinking it…

  • communism@lemmy.ml
    link
    fedilink
    arrow-up
    1
    ·
    2 days ago

    I don’t have many so I just put them all in ~/own_repos (as opposes to another directory I have for other people’s git repos). Maybe if/when I have more projects I’ll have a more elaborate system.

  • TimeSquirrel@kbin.melroy.org
    link
    fedilink
    arrow-up
    4
    ·
    2 days ago

    As an addition to your post, I’m also in the process of learning C/C++, and I’m curious also how others arrange their actual project files and include directories. Like, for example, if there’s a bunch of classes having to do with UI elements, do you just group them each under their own file all in their own directory? I’ve also seen projects where everything was just thrown into the top level directory, both headers and implementation files together in a giant pile of source files.

  • Matt@lemmy.ml
    link
    fedilink
    arrow-up
    3
    ·
    1 day ago

    Whatever Cargo generates for me. If I use workspaces, then I put the subprojects to the root of the directory.

  • moreeni@lemm.ee
    link
    fedilink
    arrow-up
    5
    ·
    1 day ago

    I have a dedicated directory with subdirectories for each project and that’s it

  • MXX53@programming.dev
    link
    fedilink
    arrow-up
    2
    ·
    14 hours ago

    I have a folder for my projects on root and within those projects I have my GitHub repos all contained within their own directory named the same as the project.

    If I am learning something, I have a folder for the topic I am learning, and a logseq file with all of my notes. Then I have folders for my book references, one for video or audio references, and then a folder for my practice projects.

  • Nibodhika@lemmy.world
    link
    fedilink
    arrow-up
    2
    ·
    1 day ago

    On my personal computer ~/Projects/<name>, you need to remember that real-life is not like college, you won’t be working on a new project every week. If you have more stuff than you can manage like this, you’ve bitten more than you can chew.

    On my work computer it’s a bit more complex, because I have to work with other people’s projects as well, so I have a ~/Work folder and in it several folders by type of stuff, e.g. ops for operational stuff such as scripts to deploy stuff or grant permissions, code for servers (and client) code, etc. Also if I’m working on something specific that requires multiple repos I create a folder for that project with the repos inside.

  • linearchaos@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 day ago

    I keep a root folder. On Windows it’s in c:\something on Linux it’s in /something

    Under there I’ve got projects organized by language. This helps me organize nix shells and venvs.

    Syncthing keeps the code bases and synced between multiple computers

    I don’t separate work from home because they don’t live in the same realm.

    Only home stuff in the syncthing.

  • telepresence@discuss.tchncs.de
    link
    fedilink
    arrow-up
    3
    ·
    2 days ago

    i use coding/languagename/project, where most if not all projects are git repos. so, coding/python/shira, coding/java/datetime examples i have some wildcards for the languages, most of my serverside js stuff lives in coding/node-deno and most of my fullstack webdev stuff lives in coding/webdev

    i used to have the coding directory on a hdd, but moving it to an ssd helped a lot when installing things made with node, among other things.

    • faultypidgeon@programming.dev
      link
      fedilink
      arrow-up
      4
      ·
      2 days ago

      I used to do this, but imho the used language is hardly a useful index. When does it happen that you want to see everything written python? For me that’s never.

      Also where do you put multi-language projects? Like, go backend with typescript frontend or whatever.

  • fruitycoder@sh.itjust.works
    link
    fedilink
    arrow-up
    4
    ·
    2 days ago

    ${HOME}/Projects/(Personal|Work)/<project name>

    If either folder gets to busy I start to create projects Meta folders that normally corospond with a gitlab group.

    I just blow out the folders with a good ol rm -rf ./ And git pull if I want to mess with it again.

  • Katzenmann@feddit.org
    link
    fedilink
    English
    arrow-up
    1
    ·
    14 hours ago

    I just have all my active projects under $HOME/projects/ and projects from other people in $HOME/bin/

  • Daniel Quinn@lemmy.ca
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    edit-2
    2 days ago
    • Keep everything in an external git service. You can use third party services like Codeberg, GitLab, or GitHub, or host your own on your NAS.
    • When you’re not working on a project and don’t think you’ll need to reference it for a while, just delete it from your laptop. The code always lives in git anyway.

    In terms of local storage, I usually have everything in ~/projects/project-name, and I don’t have tiny file size limits because I don’t use FAT32 filesystems — that’s the default filesystem you usually get on USB sticks and external hard drives you buy. You have to format those drives to something like EXT4 (Linux) or NTFS (Windows) or you get stuck with FAT32 which has 2gb file sizes.

    • stewi@sh.itjust.works
      link
      fedilink
      arrow-up
      3
      ·
      2 days ago

      You’ve forgotten about exFAT my dude. Nothing uses FAT32 anymore. All your usb drives will be exFAT.

      • Daniel Quinn@lemmy.ca
        link
        fedilink
        English
        arrow-up
        3
        ·
        2 days ago

        That’s not been my experience. Lots of drives I’ve bought have been FAT32 out of the box.

    • N0x0n@lemmy.ml
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      2 days ago

      What about exFAT? It overcomes FAT32’s limitation and is nearly readable on every OS and has way higher file size limits.

      Edit: In case of external storage like USB/hard drive.

      • Daniel Quinn@lemmy.ca
        link
        fedilink
        English
        arrow-up
        5
        ·
        2 days ago

        ExFAT is good for portable devices, but if you’re working with something internally, there’s no reason not to use EXT4 or NTFS.

  • Dragonish@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    7
    ·
    2 days ago

    Hosting my own git server on my NAS made my life easier and better due to the new freedoms it offers. Backups are centralized, and I have all the space i need to keep any interesting code safe. I am using forgjeo now and highly recommend it. You can also use other front ends (or none and just ssh/filesystem) but forgjeo gives me artifacts (ie docker registry), code search, LFS, and more. With my own git server, my local filesystem only has what I am working on recently (or as my workstation space allows). My home folder has a folder for each version control system (git, pijul, svn). Inside of these i have 2 sub folders: <domain>/<repo name>

    Some examples of different domains are: open, work, personal, dragonish. I do not separate what forge or remote service in the filesystem, this is a persona boundary.

    I use git remote names and branches in each repo to handle what software forge and any upstream/maintainers i need to work with. As an example my work repos only get pushed to my work server (ie, only 1 git remote named origin set to my work’s server), but my open ones will go to forgjeo and github (i setup 2 git remotes, origin and github. origin in this domain goes to my forgjeo). If i have a need i go into some more git branching strategy which I do find has helped me over my life, but I think I am overthinking this post now! keep it all simple enough for what your needs are.

    • Jay@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 days ago

      I also use a NAS as my primary git server, although I’ve never heard of Forgejo before, but it sounds really interesting and definitely something that I may look into. It sounds really handy to have on hand, especially with code search and LFS. So thanks for introducing that!

    • crimsonpoodle@pawb.socialOP
      link
      fedilink
      arrow-up
      1
      ·
      1 day ago

      Def going to check this out— I think it will solve my biggest problem with GitHub which is just the number of large files I have to deal with— overthought comments are the best comments- thanks!

  • faultypidgeon@programming.dev
    link
    fedilink
    arrow-up
    6
    ·
    2 days ago

    I push every project I work on right away to my gitea instance. If I expect not to work on something for some time I just delete the local copy.

    When I change devices or hit file size limits, I’ll compress and send things to my NAS.

    Well, that sounds inconvenient.

    • crimsonpoodle@pawb.socialOP
      link
      fedilink
      arrow-up
      3
      ·
      1 day ago

      Yeah, I really should start using Git for everything, but I’ve been working with a lot of large datasets recently (mostly EEG data). A big part of improving accuracy comes from cleaning the data, which is huge and takes a while to process. I could set up a local Git server to keep track of everything or just save the base data files and regenerate as needed, but on my current setup, that process can take anywhere from 2-6 hours depending on the task. So for now, I’ve just been managing everything locally to save time.