Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
308 changes: 154 additions & 154 deletions .github/actions/lava_job_render/action.yml
Original file line number Diff line number Diff line change
@@ -1,154 +1,154 @@
name: Test Action
inputs:
docker_image:
description: Docker image
required: true
default: kmake-image:latest

runs:
using: "composite"
steps:
- name: Process presigned_urls.json
id: process_urls
uses: actions/github-script@v7
with:
script: |
const fs = require('fs');
const p = require('path');
// Helper function to find URL by filename
function findUrlByFilename(filename) {
for (const [path, url] of Object.entries(data)) {
if (path.endsWith(filename)) {
return url;
}
}
return null;
}
const filePath = p.join(process.env.GITHUB_WORKSPACE, 'presigned_urls.json');
if (fs.existsSync(filePath)) {
console.log("File exists");
} else {
console.log("File does not exist");
core.setFailed(`File not found: ${filePath}`);
}
// Read the JSON file
const data = JSON.parse(fs.readFileSync(filePath, 'utf-8'));
// Extract URLs into variables
const modulesTarUrl = findUrlByFilename('modules.tar.xz');
const imageUrl = findUrlByFilename('Image');
const vmlinuxUrl = findUrlByFilename('vmlinux');
const dtbUrl = findUrlByFilename('qcs6490-rb3gen2.dtb');
// Set outputs
core.setOutput('modules_url', modulesTarUrl);
core.setOutput('image_url', imageUrl);
core.setOutput('vmlinux_url', vmlinuxUrl);
core.setOutput('dtb_url', dtbUrl);
console.log(`Modules URL: ${modulesTarUrl}`);
console.log(`Image URL: ${imageUrl}`);
console.log(`Vmlinux URL: ${vmlinuxUrl}`);
console.log(`Dtb URL: ${dtbUrl}`);

- name: Create metadata.json
id: create_metadata
shell: bash
run: |
echo "Creating job definition"
# Create the job definition using the processed URLs
cd ../job_render
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e dtb_url="${{ steps.process_urls.outputs.dtb_url }}" \
${{ inputs.docker_image }} \
jq '.artifacts["dtbs/qcom/qcs6490-rb3gen2.dtb"] = env.dtb_url' data/metadata.json > temp.json && mv temp.json data/metadata.json

- name: Upload metadata.json
id: upload_metadata
uses: ./.github/actions/aws_s3_helper
with:
local_file: ../job_render/data/metadata.json
s3_bucket: qli-prd-kernel-gh-artifacts
mode: single-upload

- name: Create template json
shell: bash
run: |
echo "Creating job definition"
metadata_url="${{ steps.upload_metadata.outputs.presigned_url }}"
vmlinux_url="${{ steps.process_urls.outputs.vmlinux_url }}"
image_url="${{ steps.process_urls.outputs.image_url }}"
modules_url="${{ steps.process_urls.outputs.modules_url }}"
# Create the job definition using the processed URLs
cd ../job_render
# using metadata_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e metadata_url="$metadata_url" \
${{ inputs.docker_image }} \
jq '.artifacts.metadata = env.metadata_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
# using image_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e image_url="$image_url" \
${{ inputs.docker_image }} \
jq '.artifacts.kernel = env.image_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
# using vmlinux_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e vmlinux_url="$vmlinux_url" \
${{ inputs.docker_image }} \
jq '.artifacts.vmlinux = env.vmlinux_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
# using modules_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e modules_url="$modules_url" \
${{ inputs.docker_image }} \
jq '.artifacts.modules = env.modules_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json

- name: Update firmware and ramdisk
shell: bash
run: |
cd ../job_render
ramdisk_url="$(aws s3 presign s3://qli-prd-kernel-gh-artifacts/meta-qcom/initramfs-kerneltest-full-image-qcom-armv8a.cpio.gz --expires 7600)"
firmware_url="$(aws s3 presign s3://qli-prd-kernel-gh-artifacts/meta-qcom/initramfs-firmware-rb3gen2-image-qcom-armv8a.cpio.gz --expires 7600)"
# using ramdisk_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e ramdisk_url="$ramdisk_url" \
${{ inputs.docker_image }} \
jq '.artifacts.ramdisk = env.ramdisk_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json

# using firmware_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e firmware_url="$firmware_url" \
${{ inputs.docker_image }} \
jq '.artifacts.firmware = env.firmware_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json

- name: Create lava_job_definition
shell: bash
run: |
cd ../job_render
mkdir renders
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
${{ inputs.docker_image }} \
sh -c 'export BOOT_METHOD=fastboot && \
export TARGET=qcs6490-rb3gen2 && \
export TARGET_DTB=qcs6490-rb3gen2 && \
python3 lava_Job_definition_generator.py --localjson ./data/cloudData.json'
name: Test Action
inputs:
docker_image:
description: Docker image
required: true
default: kmake-image:latest
runs:
using: "composite"
steps:
- name: Process presigned_urls.json
id: process_urls
uses: actions/github-script@v7
with:
script: |
const fs = require('fs');
const p = require('path');
// Helper function to find URL by filename
function findUrlByFilename(filename) {
for (const [path, url] of Object.entries(data)) {
if (path.endsWith(filename)) {
return url;
}
}
return null;
}
const filePath = p.join(process.env.GITHUB_WORKSPACE, 'presigned_urls.json');
if (fs.existsSync(filePath)) {
console.log("File exists");
} else {
console.log("File does not exist");
core.setFailed(`File not found: ${filePath}`);
}
// Read the JSON file
const data = JSON.parse(fs.readFileSync(filePath, 'utf-8'));
// Extract URLs into variables
const modulesTarUrl = findUrlByFilename('modules.tar.xz');
const imageUrl = findUrlByFilename('Image');
const vmlinuxUrl = findUrlByFilename('vmlinux');
const dtbUrl = findUrlByFilename('qcs6490-rb3gen2.dtb');
// Set outputs
core.setOutput('modules_url', modulesTarUrl);
core.setOutput('image_url', imageUrl);
core.setOutput('vmlinux_url', vmlinuxUrl);
core.setOutput('dtb_url', dtbUrl);
console.log(`Modules URL: ${modulesTarUrl}`);
console.log(`Image URL: ${imageUrl}`);
console.log(`Vmlinux URL: ${vmlinuxUrl}`);
console.log(`Dtb URL: ${dtbUrl}`);
- name: Create metadata.json
id: create_metadata
shell: bash
run: |
echo "Creating job definition"
# Create the job definition using the processed URLs
cd ../job_render
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e dtb_url="${{ steps.process_urls.outputs.dtb_url }}" \
${{ inputs.docker_image }} \
jq '.artifacts["dtbs/qcom/qcs6490-rb3gen2.dtb"] = env.dtb_url' data/metadata.json > temp.json && mv temp.json data/metadata.json
- name: Upload metadata.json
id: upload_metadata
uses: ./.github/actions/aws_s3_helper
with:
local_file: ../job_render/data/metadata.json
s3_bucket: qli-prd-kernel-gh-artifacts
mode: single-upload
- name: Create template json
shell: bash
run: |
echo "Creating job definition"
metadata_url="${{ steps.upload_metadata.outputs.presigned_url }}"
vmlinux_url="${{ steps.process_urls.outputs.vmlinux_url }}"
image_url="${{ steps.process_urls.outputs.image_url }}"
modules_url="${{ steps.process_urls.outputs.modules_url }}"
# Create the job definition using the processed URLs
cd ../job_render
# using metadata_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e metadata_url="$metadata_url" \
${{ inputs.docker_image }} \
jq '.artifacts.metadata = env.metadata_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
# using image_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e image_url="$image_url" \
${{ inputs.docker_image }} \
jq '.artifacts.kernel = env.image_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
# using vmlinux_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e vmlinux_url="$vmlinux_url" \
${{ inputs.docker_image }} \
jq '.artifacts.vmlinux = env.vmlinux_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
# using modules_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e modules_url="$modules_url" \
${{ inputs.docker_image }} \
jq '.artifacts.modules = env.modules_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
- name: Update firmware and ramdisk
shell: bash
run: |
cd ../job_render
ramdisk_url="$(aws s3 presign s3://qli-prd-kernel-gh-artifacts/meta-qcom/initramfs-kerneltest-full-image-qcom-armv8a.cpio.gz --expires 7600)"
firmware_url="$(aws s3 presign s3://qli-prd-kernel-gh-artifacts/meta-qcom/initramfs-firmware-rb3gen2-image-qcom-armv8a.cpio.gz --expires 7600)"
# using ramdisk_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e ramdisk_url="$ramdisk_url" \
${{ inputs.docker_image }} \
jq '.artifacts.ramdisk = env.ramdisk_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
# using firmware_url
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
-e firmware_url="$firmware_url" \
${{ inputs.docker_image }} \
jq '.artifacts.firmware = env.firmware_url' data/cloudData.json > temp.json && mv temp.json data/cloudData.json
- name: Create lava_job_definition
shell: bash
run: |
cd ../job_render
mkdir renders
docker run -i --rm \
--user "$(id -u):$(id -g)" \
--workdir="$PWD" \
-v "$(dirname "$PWD")":"$(dirname "$PWD")" \
${{ inputs.docker_image }} \
sh -c 'export BOOT_METHOD=fastboot && \
export TARGET=qcs6490-rb3gen2 && \
export TARGET_DTB=qcs6490-rb3gen2 && \
python3 lava_Job_definition_generator.py --localjson ./data/cloudData.json --KernelCI_PreMerge'
Loading
Loading