Mulesoft MCD-Level-1 Exam Questions

229 Questions


Updation Date : 13-Jan-2026



Mulesoft MCD-Level-1 exam questions feature realistic, exam-like questions that cover all key topics with detailed explanations. You’ll identify your strengths and weaknesses, allowing you to focus your study efforts effectively. By practicing with our MCD-Level-1 practice test, you’ll gain the knowledge, speed, and confidence needed to pass the Mulesoft exam on your first attempt.

Why leave your success to chance? Our Mulesoft MCD-Level-1 dumps are your ultimate guide to passing the exam on your first try!

An API implementation has been deployed to CloudHub and now needs to be governed. IT
will not allocate additional vCore for a new Mule application to act as an API proxy.
Whatis the next step to preseive the current vCore usage, but still allow the Mule
application to be managed by API Manager?


A.

Register the same API implementation in Runtime Manager to connect to API Manager


B.

Modify the API implementation to useauto-discovery to register with API Manager


C.

Upload the Mule application's JAR file to the API instance in API Manager


D.

Deploy the same API implementation behind a VPC and configure the VPC to connect
to API Manager





B.
  

Modify the API implementation to useauto-discovery to register with API Manager



What MuleSoft API-led connectivity layer is intended to expose part of a backend database without business logic?


A. Data layer


B. Process layer


C. Experience layer


D. System layer





D.
  System layer

Explanation:
Correct answer is System layer
System APIs provide a means for insulating the data consumers from the complexity or changes to the underlying backend systems.
MuleSoft recommends three-layered approach to API-led connectivity, highlighting the three layers:
* System APIs
* Process APIs
* Experience APIs
System APIs are the core systems of record underlying core systems of record (e.g. ERPs, key customer and billing systems, databases, etc.). Process APIs allow you to define a common process which the organization can share, and these APIs perform specific functions, provide access to non-central data, and may be built by either Central IT or Line of Business IT. And finally, the Experience APIs are the means by which data can be reconfigured so that it is most easily consumed by its intended audience, all from a common data source.
The three-layered structure allows for a seamless flow of data from systems of record to new experiences, and allows for reusability of assets rather than point to point connections.
This approach provides a distributed and tailored approach to architecture, greater flexibility through loose coupling, and deeper operational visibility into what is being built.

An SLA based policy has been enabled in API Manager. What is the next step to configure
the API proxy to enforce the newSLA policy?


A.

Add new property placeholders and redeploy the API proxy


B.

Add new environment variables and restart the API proxy


C.

Restart the API proxy to clear the API policy cache


D.

Add required headers to the RAML specification and redeploy the newAPI proxy





D.
  

Add required headers to the RAML specification and redeploy the newAPI proxy



A mule project contains MySQL database dependency. The project is exported from Anypoint Studio so that it can be deployed to Cloudhub. What export options needs to be selected to create the smallest deployable archive that will successfully deploy to Cloudhub?


A. Select both the options 1) Attach project sources 2) Include project module and dependencies


B. No need to select any of the below options 1) Attach project sources 2) Include project module and dependencies


C. Select only below option 2) Include project module and dependencies


D. Select only below option 1) Attach project sources





C.
  Select only below option 2) Include project module and dependencies

Explanation:
You can choose Attach Project Sources to include metadata that Studio requires to reimport the deployable file as an open Mule project into your workspace. You must keep the Attach Project Sources option selected to be able to import the packaged JAR file back into a Studio workspace. But requirement here is to create smallest deployable archive that will successfully deploy to Cloudhub. Hence we can ignore this option.
We need to select Include project module and dependencies.
As actual modules and external dependencies required to run the Mule application in a Mule runtime engine.

Which of the below activity doesn't support parallel execution?


A. Scatter-Gather Router


B. First Successful Router


C. Parallel For Each


D. Batch job





C.
  Parallel For Each

An On Table Row Database listener retrieves data from a table that contains record_id, an
increasing numerical column.
How should the listener be configured so it retrieves new rows at most one time?


A.

Set the target tostore the last retrieved record_id value


B.

Set the ObjectStore to store the last retrieved record_id value


C.

Set the target to the record_id column


D.

Set the watermark column to the record id column





D.
  

Set the watermark column to the record id column



Refer to the exhibits.

The<when> expression for the Choice router needs to be written.
What is a valid <when> expression to route Mule events to the non-default flow?


A.

#['MuleSoft' == paytoad.company]


B.

#[ company = "MuleSoft" ]


C.

#[ if( company = "MuleSoft") ]


D.

#[ if('MuleSoff == payload.company) ]





A.
  

#['MuleSoft' == paytoad.company]



A Batch Job scope has five batch steps. An event processor throws an error in the second batch step because the input data is incomplete. What is the default behavior of the batch job after the error is thrown?


A. All processing of the batch job stops.


B. Event processing continues to the next batch step.


C. Error is ignored


D. Batch is retried





A.
  All processing of the batch job stops.

Explanation:
In case of an error , batch job completes in flight steps and stops further processing.
MuleSoft Doc Ref : Handling Errors During Batch Job | MuleSoft Documentation The default is all processing will stop but we can change it by Max Failed Record field.
General -> Max Failed Records: Mule has three options for handling a record-level error:
Finish processing, Continue processing and Continue processing until the batch job accumulates a maximum number of failed records. This behavior can be controlled by Max Failed Records.
The default value is Zero which corresponds to Finish processing.
The value -1, corresponds to Continue processing.
The value +ve integer, corresponds to Continue processing until the batch job accumulates a maximum number of failed records.


Page 1 out of 29 Pages