Distributed Compiling and Operations Manager for RockY Linux !
What about doing some High Performance Computing (HPC) all along the process of building from zero Rocky Enterprise Linux ?
A lot of people and organizations has offered computing resources to help Rocky Linux be lifted off asap ! So, why not combine all that in one shot, helping development, testing, integration, packages compatibility checks as well other fundamental tasks and get Rocky ready for action any time soon ?
Under the hood tools
distcc
icecream
bistro
Follow steps on https://gitlab.com/SkipGrube/RockyLinux/-/blob/master/Build_Steps.md
Those are some desirable features to be reached :
[ ] Design interactions with Koji and Mock
[ ] Prepare job queues
[ ] Connect to agents
[ ] Check agents availability
[ ] Signal agents for green light
[ ] Check the jobs queue for priority
[ ] Start jobs distribution
[ ] Monitor jobs execution
[ ] Check job finishing
[ ] Update queue
[ ] Notify CI/CD server of the conclusion
Discomy will come in two flavors : discomy-client
and discomy-server
, each having specifications for receiving jobs and passing away to the servers for heavy lift and bulk work (compiling, checking, sanitization etc).
To acomplish all that, the following packages will be the core of discomy
:
job_listener
Through web hooks, catch all work done in remote code repo's, read the pipeline specifications, and start to distribute the incoming jobs to the queue.
job_queue
Get the building steps onto the queue, with specifications imported from the CI/CD pipeline, caches the entries and pull the ongoing tasks ahead the chain.
host_scheduler
Reads the entries from job_queue
, searchs for free hosts, allocates resources (make isolations with cgroups and fakeroot-like processes), prepare the background to process incoming jobs and waits for the green light.
host_agregator
Syncs all waiting and executing jobs, add lookups for them, combine their outputs, creates namespaces (more like grouping those jobs to be spawned with specific tasks in hand).
host_executor
For both client and server tasks, host_executor
takes all the jobs in the queue and executes.
queue_executor
The minor processing unit for all kind of jobs through the chain. Runs concurrently in background (hello goroutines :D), for both client and server, publishes its outputs (job signal changes from waiting
to done
).
final_agregator
Looks for all finished jobs into the chain, determines what to do next based on the job context, priority and type (compress the final file or package, starts another job, pushes the results to some repo, make some checksum etc).
For dealing with config files, and interfacing with API's, two more packages are in place :
config_manager
Handle all configurations necessary to the execution of virtual machines, web hooks keys, authentication tokens and so on.
api
External tools necessary for accomplishing a lot of steps, as well libraries for infrastructure provisioning, configuration, delegation and so on will handled here.
Other non-functional requirements in turn to make possible :
Possible batch execution - ex. package sanitization functions, compatibility checking.
Users apply for a subscription to include their computing capacity, filling a form with the info necessary to provisioning ("disco playlists"). After then, can download provisioning playbooks and manifests, or just configure by hand and insert the info of compiling grid - approved nodes will receive keys, security tokens and certificates to join the "working fleet" (in a near future, a web UI (wuimy
) will allow to include emails and users profiles so all of this will be automated using trusty process).
Calculate the theoretical peak to finish the compiling step, and give an estimate of how much computing and storage is necessary for all - based on the number of available hosts, CPU I/O, storage, network paths.
Run sync with repos and mirror after finishing, checking integrity of past jobs.
To test the basics, and check each step of desired features, virtual envs as well network of hosts is created, jobs assigned to it and the discomy
monitoring tool (moniscomy
or monitoring for discomy).
For the initial steps, Vermin can be used to spawn some virtual machines, inject discomy through an installation and configuration script (install.sh and configure.sh), test the outputs and get the final results.
A RPM package creation will be placed, and the steps for building versions of the same package (maybe following git-flow
) to get a "stable" branch of the package.
In more fine grained way, package checks can be done too, to test tool capability, limitations and bottlenecks.
To make some contribution, just fork the repo, do your stuff, and make a PR so we can align our work.