pedroalvesbatista / discomy

1 stars 0 forks source link

Discomy

Distributed Compiling and Operations Manager for RockY Linux !

What about doing some High Performance Computing (HPC) all along the process of building from zero Rocky Enterprise Linux ?

A lot of people and organizations has offered computing resources to help Rocky Linux be lifted off asap ! So, why not combine all that in one shot, helping development, testing, integration, packages compatibility checks as well other fundamental tasks and get Rocky ready for action any time soon ?

Under the hood tools

distcc

icecream

bistro

How it works ?

Follow steps on https://gitlab.com/SkipGrube/RockyLinux/-/blob/master/Build_Steps.md

Those are some desirable features to be reached :

Discomy will come in two flavors : discomy-client and discomy-server, each having specifications for receiving jobs and passing away to the servers for heavy lift and bulk work (compiling, checking, sanitization etc).

To acomplish all that, the following packages will be the core of discomy :

job_listener

Through web hooks, catch all work done in remote code repo's, read the pipeline specifications, and start to distribute the incoming jobs to the queue.

job_queue

Get the building steps onto the queue, with specifications imported from the CI/CD pipeline, caches the entries and pull the ongoing tasks ahead the chain.

host_scheduler

Reads the entries from job_queue, searchs for free hosts, allocates resources (make isolations with cgroups and fakeroot-like processes), prepare the background to process incoming jobs and waits for the green light.

host_agregator

Syncs all waiting and executing jobs, add lookups for them, combine their outputs, creates namespaces (more like grouping those jobs to be spawned with specific tasks in hand).

host_executor

For both client and server tasks, host_executor takes all the jobs in the queue and executes.

queue_executor

The minor processing unit for all kind of jobs through the chain. Runs concurrently in background (hello goroutines :D), for both client and server, publishes its outputs (job signal changes from waiting to done).

final_agregator

Looks for all finished jobs into the chain, determines what to do next based on the job context, priority and type (compress the final file or package, starts another job, pushes the results to some repo, make some checksum etc).

For dealing with config files, and interfacing with API's, two more packages are in place :

config_manager

Handle all configurations necessary to the execution of virtual machines, web hooks keys, authentication tokens and so on.

api

External tools necessary for accomplishing a lot of steps, as well libraries for infrastructure provisioning, configuration, delegation and so on will handled here.


Other non-functional requirements in turn to make possible :

Testing sandbox

To test the basics, and check each step of desired features, virtual envs as well network of hosts is created, jobs assigned to it and the discomy monitoring tool (moniscomy or monitoring for discomy).

For the initial steps, Vermin can be used to spawn some virtual machines, inject discomy through an installation and configuration script (install.sh and configure.sh), test the outputs and get the final results.

A RPM package creation will be placed, and the steps for building versions of the same package (maybe following git-flow) to get a "stable" branch of the package.

In more fine grained way, package checks can be done too, to test tool capability, limitations and bottlenecks.

Contributing

To make some contribution, just fork the repo, do your stuff, and make a PR so we can align our work.