Q7 & Q9 wrong in CKA Mock Exam #4?

Hey, I’m playing your mock exams for CKA, but I have problems with questions 7 and 9 of the
Lab: CKA Mock Exam 4

Question 7:
The solution provided by you is:

Update `service-cka25-arch.sh` script:
student-node ~ ➜ vi service-cka25-arch.sh

Add below command in it:

kubectl --context cluster1 get service service-cka25-arch -o jsonpath='{.spec.ports[0].targetPort}'

My question was the following, which is basically the same, but marked as wrong:

student-node ~ ➜  cat /root/service-cka25-arch.sh
#!/bin/bash
kubectl get svc service-cka25-arch -o jsonpath='{.spec.ports[0].targetPort}'

Question 9:
It’s also marked as wrong, but this is the YAML of the PV and PVC, so I think is correct:

student-node ~ ➜  kubectl describe pv orange-pv-cka13-trb 
Name:            orange-pv-cka13-trb
Labels:          type=local
Annotations:     pv.kubernetes.io/bound-by-controller: yes
Finalizers:      [kubernetes.io/pv-protection]
StorageClass:    
Status:          Bound
Claim:           default/orange-pvc-cka13-trb
Reclaim Policy:  Retain
Access Modes:    RWO
VolumeMode:      Filesystem
Capacity:        100Mi
Node Affinity:   <none>
Message:         
Source:
    Type:          HostPath (bare host directory volume)
    Path:          /mnt/orange-pv-cka13-trb
    HostPathType:  
Events:            <none>

student-node ~ ➜  kubectl describe pvc orange-pvc-cka13-trb 
Name:          orange-pvc-cka13-trb
Namespace:     default
StorageClass:  
Status:        Bound
Volume:        orange-pv-cka13-trb
Labels:        type=local
Annotations:   pv.kubernetes.io/bind-completed: yes
Finalizers:    [kubernetes.io/pvc-protection]
Capacity:      100Mi
Access Modes:  RWO
VolumeMode:    Filesystem
Used By:       <none>
Events:        <none>

Hi @Maiky

Regarding your query on Q7, the script you’ve posted does not define the context. The script should look as in the attached image.

With regards to Q9, the describe output of PV and PVC looks correct and the PVC is in a Bound state and should result in a pass.
If you could share the YAML format of PVC, it could have helped.

Regards.

1 Like

Sure, here you go the YAML of the PV and the PVC. I played again and it failed again.

PV:

student-node ~/app-wl03 ➜  kubectl get pv orange-pv-cka13-trb -o yaml
apiVersion: v1
kind: PersistentVolume
metadata:
  annotations:
    pv.kubernetes.io/bound-by-controller: "yes"
  creationTimestamp: "2024-11-19T05:36:10Z"
  finalizers:
  - kubernetes.io/pv-protection
  labels:
    type: local
  name: orange-pv-cka13-trb
  resourceVersion: "6414"
  uid: ec1d4152-2859-442a-b736-53a0712ccae0
spec:
  accessModes:
  - ReadWriteOnce
  capacity:
    storage: 100Mi
  claimRef:
    apiVersion: v1
    kind: PersistentVolumeClaim
    name: orange-pvc-cka13-trb
    namespace: default
    resourceVersion: "6412"
    uid: c421d520-a306-47be-80b3-f6051256fa0e
  hostPath:
    path: /mnt/orange-pv-cka13-trb
    type: ""
  persistentVolumeReclaimPolicy: Retain
  volumeMode: Filesystem
status:
  lastPhaseTransitionTime: "2024-11-19T05:38:25Z"
  phase: Bound

PVC:

student-node ~/app-wl03 ✖ kubectl get pvc orange-pvc-cka13-trb -o yaml
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  annotations:
    kubectl.kubernetes.io/last-applied-configuration: |
      {"apiVersion":"v1","kind":"PersistentVolumeClaim","metadata":{"annotations":{},"creationTimestamp":"2024-11-19T05:36:11Z","finalizers":["kubernetes.io/pvc-protection"],"labels":{"type":"local"},"name":"orange-pvc-cka13-trb","namespace":"default","resourceVersion":"6196","uid":"0bd90be9-1b2e-40e3-b20b-0c8b4775c7da"},"spec":{"accessModes":["ReadWriteOnce"],"resources":{"requests":{"storage":"100Mi"}},"volumeMode":"Filesystem","volumeName":"orange-pv-cka13-trb"},"status":{"phase":"Pending"}}
    pv.kubernetes.io/bind-completed: "yes"
  creationTimestamp: "2024-11-19T05:38:25Z"
  finalizers:
  - kubernetes.io/pvc-protection
  labels:
    type: local
  name: orange-pvc-cka13-trb
  namespace: default
  resourceVersion: "6416"
  uid: c421d520-a306-47be-80b3-f6051256fa0e
spec:
  accessModes:
  - ReadWriteOnce
  resources:
    requests:
      storage: 100Mi
  volumeMode: Filesystem
  volumeName: orange-pv-cka13-trb
status:
  accessModes:
  - ReadWriteOnce
  capacity:
    storage: 100Mi
  phase: Bound

As you can see, they’re bound:

Please, let me know if I’m wrong, I’d appreciate it.

This looks correct to me and should result in a pass.
This Question requires updating the PVC requested storage from 150Mi to a value equal to or less than 100Mi.

Can you please try it afresh?
Please note you need to perform the action on PVC on cluster1

kubectl config use-context cluster1

Regards.