Uploaded image for project: 'Spark Connector'
  1. Spark Connector
  2. SPARK-399

Spark connector 10.1.1 not supporting microbatch processing

    • Hide

      1. What would you like to communicate to the user about this feature?
      2. Would you like the user to see examples of the syntax and/or executable code and its output?
      3. Which versions of the driver/connector does this apply to?

      Show
      1. What would you like to communicate to the user about this feature? 2. Would you like the user to see examples of the syntax and/or executable code and its output? 3. Which versions of the driver/connector does this apply to?

      Hi Team,

      We are trying out the AWS Glue --> Pyspark connector for MongoDB Atlas.  While trying out the spark streaming, we are getting the following error.

      Error: Data Source com.mongodb.spark.sql.connector.MongoTableProvider does not support microbatch processing.

      Let us know, if you need any further details.

      Appreciate your support.

      Attached is the error screenshot, the Pyspark code used for streaming, and a high-level flow diag.

        1. Microbatch_Join_Notworking-1.png
          65 kB
          Babu Srinivasan
        2. MongoGlueJoin_MicroBatch_NotWorking-1.py
          3 kB
          Babu Srinivasan
        3. Spark_Streaming_Join_MicroBatch_NotWorking-1.png
          597 kB
          Babu Srinivasan

            Assignee:
            prakul.agarwal@mongodb.com Prakul Agarwal
            Reporter:
            babu.srinivasan@mongodb.com Babu Srinivasan
            Votes:
            0 Vote for this issue
            Watchers:
            2 Start watching this issue

              Created:
              Updated:
              Resolved: