Hello,
We are using Apache Flink 1.10.1 version. During our security scans following issues are reported by our scan tool. Please let us know your comments on these dependency vulnerabilities. Thanks, Suchithra -----Original Message----- From: [hidden email] <[hidden email]> On Behalf Of Apache Security Team Sent: Thursday, August 6, 2020 1:08 PM To: V N, Suchithra (Nokia - IN/Bangalore) <[hidden email]> Cc: Jash, Shaswata (Nokia - IN/Bangalore) <[hidden email]>; Prabhala, Anuradha (Nokia - IN/Bangalore) <[hidden email]>; Badagandi, Srinivas B. (Nokia - IN/Bangalore) <[hidden email]> Subject: Re: Security vulnerabilities with Apache Flink 1.10.1 version Hi, Outdated dependencies are not always security issues. A project would only be affected if a dependency was used in such a way that the affected underlying code is used and the vulnerabilities were exposed. We typically get reports sent to us from scanning tools that looks at dependencies out of context on how they are actually used in the projects. As such we reject these reports and suggest you either a) show how the product is affected by the dependency vulnerabilities, or b) simply mention this as a normal bug report to that project. Since dependency vulnerabilities are quite public, there is no need to use this private reporting mechanism for them. Regards, Mark On Thu, Aug 6, 2020 at 6:04 AM V N, Suchithra (Nokia - IN/Bangalore) <[hidden email]> wrote: > > Hello, > > > > We are using Apache Flink 1.10.1 version. During our security scans following issues are reported by our scan tool. > > > > 1.Package : log4j-1.2.17 > > Severity: CRITICAL > > Fix version: 2.8.2 > > > > Description: > > Apache Log4j contains a flaw that is triggered as the SocketServer class accepts log data from untrusted network traffic, which it then insecurely deserializes. This may allow a remote attacker to potentially execute arbitrary code. > > > > Path: > > /opt/flink/lib/log4j-1.2.17.jar > > /opt/flink/bin/bash-java-utils.jar:log4j > > > > References: > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2019-17571 > > https://seclists.org/oss-sec/2019/q4/167 > > https://logging.apache.org/log4j/1.2/ > > > > 2.Package: guava-14.0.1 > > Severity: HIGH > > Fix version: 25.0, 24.1.1 > > > > Description: > > Google Guava contains a flaw in the CompoundOrdering_CustomFieldSerializer::deserialize() function in com/google/common/collect/CompoundOrdering_CustomFieldSerializer.java that is triggered when deserializing Java objects. With specially crafted serialized data, a context-dependent can exhaust available memory, resulting in a denial of service. > > > > References: > > https://github.com/google/guava/wiki/CVE-2018-10237 > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237 > > > > 3.Package: guava-14.0.1,18.0 > > Severity: HIGH > > Fix version: 25.0, 24.1.1 > > > > Description: > > Google Guava contains a flaw in the AtomicDoubleArray::readObject() function in com/google/common/util/concurrent/AtomicDoubleArray.java that is triggered when deserializing Java objects. With specially crafted serialized data, a context-dependent can cause a process linked against the library to exhaust available memory. > > > > References: > > https://github.com/google/guava/wiki/CVE-2018-10237 > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237 > > > > 4. Package: guava-19.0 > > Severity: HIGH > > Fix version: 25.0, 24.1.1, 23.6.1 > > > > Description: > > Google Guava contains a flaw in the CompoundOrdering_CustomFieldSerializer::deserialize() function in com/google/common/collect/CompoundOrdering_CustomFieldSerializer.java that is triggered when deserializing Java objects. With specially crafted serialized data, a context-dependent can exhaust available memory, resulting in a denial of service. > > > > References: > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237 > > https://github.com/google/guava/wiki/CVE-2018-10237 > > > > Path: > > /opt/flink/lib/flink-table-blink_2.11-1.10.1.jar:guava > > /opt/flink/lib/flink-table_2.11-1.10.1.jar:guava > > /opt/flink/lib/flink-dist_2.11-1.10.1.jar:guava > > /opt/flink/examples/streaming/Twitter.jar:guava > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:guava > > /opt/flink/lib/flink-table-blink_2.11-1.10.1.jar:guava > > /opt/flink/lib/flink-table_2.11-1.10.1.jar:guava > > /opt/flink/lib/flink-dist_2.11-1.10.1.jar:guava > > /opt/flink/examples/streaming/Twitter.jar:guava > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:guava > > > > 5.Package: commons_io-2.4 > > Severity: HIGH > > Fix version: 2.5 > > > > Description: > > Apache Commons IO contains a flaw that is due to the program failing to restrict which class can be serialized. This may allow a remote attacker to execute arbitrary Java code via deserialization methods. > > > > Path: > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:commons-io > > /opt/flink/lib/flink-dist_2.11-1.10.1.jar:commons-io > > > > References: > > https://issues.apache.org/jira/browse/IO-487 > > http://commons.apache.org/proper/commons-io/ > > > > 6.Package: commons_beanutils-1.9.3 > > Severity: HIGH > > Fix version: 1.9.3 > > > > Description: > > Commons BeanUtils contains a flaw that is due to it failing to restrict the setting of Class Loader attributes via the class attribute of an ActionForm object. This may allow a remote attacker to manipulate the loaded Classloader and execute arbitrary code. > > > > Path: > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:commons-beanut > ils > > > > References: > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2014-0114 > > https://seclists.org/bugtraq/2014/Apr/177 > > > > 7.Package: hadoop-2.6.5 > > Severity: HIGH > > Fix version: 3.1.1, 3.0.3, 2.9.2, 2.8.5 > > > > Description: > > Apache Hadoop contains a flaw that allows traversing outside of a restricted path. The issue is due to the FileUtil::unpackEntries() functions in fs/FileUtil.java not properly sanitizing user input, specifically path traversal style attacks (e.g. '../') supplied via filenames. With a specially crafted ZIP archive, a context-dependent attacker can write to arbitrary files. > > > > References: > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8009 > > https://github.com/snyk/zip-slip-vulnerability > > > > 8.Package: hadoop-2.6.5 > > Severity: HIGH > > Fix version: 3.1.1, 3.0.3, 2.9.2, 2.8.5 > > > > Description: > > Apache Hadoop contains a flaw that allows traversing outside of a restricted path. The issue is due to the FileUtil::unZip() functions in fs/FileUtil.java not properly sanitizing user input, specifically path traversal style attacks (e.g. '../') supplied via filenames. With a specially crafted ZIP archive, a context-dependent attacker can write to arbitrary files. > > > > References: > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8009 > > https://github.com/snyk/zip-slip-vulnerability > > > > 9.Package: hadoop-2.6.5 > > Severity: HIGH > > Fix version: 3.1.1, 2.9.2, 2.8.5 > > > > Description: > > Apache Hadoop contains an unspecified flaw which may allow a local attacker with yarn privileges to gain elevated root privileges and execute arbitrary commands. No further details have been provided. > > > > References: > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8029 > > https://seclists.org/oss-sec/2019/q2/132 > > > > 10.Package: hadoop-2.6.5 > > Severity: HIGH > > Fix version: 2.7.7 > > > > Description: > > Apache Hadoop contains an unspecified flaw that may allow a local attacker who is able to escalate their privileges to yarn user to gain root privileges. No further details have been provided. > > > > References: > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-11766 > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2016-6811 > > > > 11.Package: hadoop-2.6.5 > > Severity: HIGH > > Fix version: 2.7.0 > > > > Description: > > Apache Hadoop contains a flaw in a servlet on the DataNode related to the HDFS namespace browsing functionality. The issue is triggered as a query parameter is not properly validated when handling NameNode information. This may allow a remote attacker to have an unspecified impact. > > > > References: > > https://cve.mitre.org/cgi-bin/cvename.cgi?name=2017-3162 > > https://seclists.org/oss-sec/2017/q2/126 > > > > 12.Package: hadoop-2.6.5 > > Severity: HIGH > > > > Description: > > In Apache Hadoop 3.1.0 to 3.1.1, 3.0.0-alpha1 to 3.0.3, 2.9.0 to 2.9.1, and 2.0.0-alpha to 2.8.4, the user/group information can be corrupted across storing in fsimage and reading back from fsimage. > > > > References: > > https://nvd.nist.gov/vuln/detail/CVE-2018-11768 > > > > Paths: > > Hadoop paths > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-co > mmon > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-cl > ient > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-ap > i > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-mapredu > ce-client-core > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-hdfs > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-common > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-auth > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-annotat > ions > > > > 13.Package: protobuf- 2.5.0, 2.6.1 > > Severity: HIGH > > > > Description: > > protobuf allows remote authenticated attackers to cause a heap-based buffer overflow. > > > > Path: > > /opt/flink/lib/flink-dist_2.11-1.10.1.jar:protobuf-java > > /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:protobuf-java > > > > References: > > https://nvd.nist.gov/vuln/detail/CVE-2015-5237 > > > > Please let us know your comments on these issues and fix plans. > > > > Regards, > > Suchithra > > |
log4j - If you don't use a Socket appender, you're good. Otherwise, you
can replace the log4j jars in lib/ with a newer version. You could also upgrade to 1.11.1 which uses log4j2. guava - We do not use Guava for serialization AFAIK. We also do not use Java serialization for records. commons-io - We do not use it for serialization. protobuf - Only used by specific components for exchanging records. If an attacker is able to introduce arbitrary data in places, then they in any case wreak havoc in all sorts of ways, be it by introducing false data invalidating your pipeline or sending obscenely large messages eating up all memory. all things contained in flink-shaded-hadoop - Please evaluate yourself whether this is problematic and for, and use a newer Hadoop version if required. Flink 1.11.0 also came with support for Hadoop 3. Flink-shaded-hadoop is no longer supported, and was just a collection of Hadoop dependencies that we package for convenience. TL;DR: Please use your own judgement for issues regarding Hadoop, and reach out to the Hadoop project for guidance. As for the remaining issues, it is unlikely that these vulnerabilities make your setup any less secure. On 06/08/2020 14:06, V N, Suchithra (Nokia - IN/Bangalore) wrote: > Hello, > > We are using Apache Flink 1.10.1 version. During our security scans following issues are reported by our scan tool. > Please let us know your comments on these dependency vulnerabilities. > > Thanks, > Suchithra > > -----Original Message----- > From: [hidden email] <[hidden email]> On Behalf Of Apache Security Team > Sent: Thursday, August 6, 2020 1:08 PM > To: V N, Suchithra (Nokia - IN/Bangalore) <[hidden email]> > Cc: Jash, Shaswata (Nokia - IN/Bangalore) <[hidden email]>; Prabhala, Anuradha (Nokia - IN/Bangalore) <[hidden email]>; Badagandi, Srinivas B. (Nokia - IN/Bangalore) <[hidden email]> > Subject: Re: Security vulnerabilities with Apache Flink 1.10.1 version > > Hi, > > Outdated dependencies are not always security issues. A project would only be affected if a dependency was used in such a way that the affected underlying code is used and the vulnerabilities were exposed. > We typically get reports sent to us from scanning tools that looks at dependencies out of context on how they are actually used in the projects. As such we reject these reports and suggest you either a) show how the product is affected by the dependency vulnerabilities, or > b) simply mention this as a normal bug report to that project. Since dependency vulnerabilities are quite public, there is no need to use this private reporting mechanism for them. > > Regards, Mark > > On Thu, Aug 6, 2020 at 6:04 AM V N, Suchithra (Nokia - IN/Bangalore) <[hidden email]> wrote: >> Hello, >> >> >> >> We are using Apache Flink 1.10.1 version. During our security scans following issues are reported by our scan tool. >> >> >> >> 1.Package : log4j-1.2.17 >> >> Severity: CRITICAL >> >> Fix version: 2.8.2 >> >> >> >> Description: >> >> Apache Log4j contains a flaw that is triggered as the SocketServer class accepts log data from untrusted network traffic, which it then insecurely deserializes. This may allow a remote attacker to potentially execute arbitrary code. >> >> >> >> Path: >> >> /opt/flink/lib/log4j-1.2.17.jar >> >> /opt/flink/bin/bash-java-utils.jar:log4j >> >> >> >> References: >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2019-17571 >> >> https://seclists.org/oss-sec/2019/q4/167 >> >> https://logging.apache.org/log4j/1.2/ >> >> >> >> 2.Package: guava-14.0.1 >> >> Severity: HIGH >> >> Fix version: 25.0, 24.1.1 >> >> >> >> Description: >> >> Google Guava contains a flaw in the CompoundOrdering_CustomFieldSerializer::deserialize() function in com/google/common/collect/CompoundOrdering_CustomFieldSerializer.java that is triggered when deserializing Java objects. With specially crafted serialized data, a context-dependent can exhaust available memory, resulting in a denial of service. >> >> >> >> References: >> >> https://github.com/google/guava/wiki/CVE-2018-10237 >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237 >> >> >> >> 3.Package: guava-14.0.1,18.0 >> >> Severity: HIGH >> >> Fix version: 25.0, 24.1.1 >> >> >> >> Description: >> >> Google Guava contains a flaw in the AtomicDoubleArray::readObject() function in com/google/common/util/concurrent/AtomicDoubleArray.java that is triggered when deserializing Java objects. With specially crafted serialized data, a context-dependent can cause a process linked against the library to exhaust available memory. >> >> >> >> References: >> >> https://github.com/google/guava/wiki/CVE-2018-10237 >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237 >> >> >> >> 4. Package: guava-19.0 >> >> Severity: HIGH >> >> Fix version: 25.0, 24.1.1, 23.6.1 >> >> >> >> Description: >> >> Google Guava contains a flaw in the CompoundOrdering_CustomFieldSerializer::deserialize() function in com/google/common/collect/CompoundOrdering_CustomFieldSerializer.java that is triggered when deserializing Java objects. With specially crafted serialized data, a context-dependent can exhaust available memory, resulting in a denial of service. >> >> >> >> References: >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237 >> >> https://github.com/google/guava/wiki/CVE-2018-10237 >> >> >> >> Path: >> >> /opt/flink/lib/flink-table-blink_2.11-1.10.1.jar:guava >> >> /opt/flink/lib/flink-table_2.11-1.10.1.jar:guava >> >> /opt/flink/lib/flink-dist_2.11-1.10.1.jar:guava >> >> /opt/flink/examples/streaming/Twitter.jar:guava >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:guava >> >> /opt/flink/lib/flink-table-blink_2.11-1.10.1.jar:guava >> >> /opt/flink/lib/flink-table_2.11-1.10.1.jar:guava >> >> /opt/flink/lib/flink-dist_2.11-1.10.1.jar:guava >> >> /opt/flink/examples/streaming/Twitter.jar:guava >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:guava >> >> >> >> 5.Package: commons_io-2.4 >> >> Severity: HIGH >> >> Fix version: 2.5 >> >> >> >> Description: >> >> Apache Commons IO contains a flaw that is due to the program failing to restrict which class can be serialized. This may allow a remote attacker to execute arbitrary Java code via deserialization methods. >> >> >> >> Path: >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:commons-io >> >> /opt/flink/lib/flink-dist_2.11-1.10.1.jar:commons-io >> >> >> >> References: >> >> https://issues.apache.org/jira/browse/IO-487 >> >> http://commons.apache.org/proper/commons-io/ >> >> >> >> 6.Package: commons_beanutils-1.9.3 >> >> Severity: HIGH >> >> Fix version: 1.9.3 >> >> >> >> Description: >> >> Commons BeanUtils contains a flaw that is due to it failing to restrict the setting of Class Loader attributes via the class attribute of an ActionForm object. This may allow a remote attacker to manipulate the loaded Classloader and execute arbitrary code. >> >> >> >> Path: >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:commons-beanut >> ils >> >> >> >> References: >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2014-0114 >> >> https://seclists.org/bugtraq/2014/Apr/177 >> >> >> >> 7.Package: hadoop-2.6.5 >> >> Severity: HIGH >> >> Fix version: 3.1.1, 3.0.3, 2.9.2, 2.8.5 >> >> >> >> Description: >> >> Apache Hadoop contains a flaw that allows traversing outside of a restricted path. The issue is due to the FileUtil::unpackEntries() functions in fs/FileUtil.java not properly sanitizing user input, specifically path traversal style attacks (e.g. '../') supplied via filenames. With a specially crafted ZIP archive, a context-dependent attacker can write to arbitrary files. >> >> >> >> References: >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8009 >> >> https://github.com/snyk/zip-slip-vulnerability >> >> >> >> 8.Package: hadoop-2.6.5 >> >> Severity: HIGH >> >> Fix version: 3.1.1, 3.0.3, 2.9.2, 2.8.5 >> >> >> >> Description: >> >> Apache Hadoop contains a flaw that allows traversing outside of a restricted path. The issue is due to the FileUtil::unZip() functions in fs/FileUtil.java not properly sanitizing user input, specifically path traversal style attacks (e.g. '../') supplied via filenames. With a specially crafted ZIP archive, a context-dependent attacker can write to arbitrary files. >> >> >> >> References: >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8009 >> >> https://github.com/snyk/zip-slip-vulnerability >> >> >> >> 9.Package: hadoop-2.6.5 >> >> Severity: HIGH >> >> Fix version: 3.1.1, 2.9.2, 2.8.5 >> >> >> >> Description: >> >> Apache Hadoop contains an unspecified flaw which may allow a local attacker with yarn privileges to gain elevated root privileges and execute arbitrary commands. No further details have been provided. >> >> >> >> References: >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8029 >> >> https://seclists.org/oss-sec/2019/q2/132 >> >> >> >> 10.Package: hadoop-2.6.5 >> >> Severity: HIGH >> >> Fix version: 2.7.7 >> >> >> >> Description: >> >> Apache Hadoop contains an unspecified flaw that may allow a local attacker who is able to escalate their privileges to yarn user to gain root privileges. No further details have been provided. >> >> >> >> References: >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-11766 >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2016-6811 >> >> >> >> 11.Package: hadoop-2.6.5 >> >> Severity: HIGH >> >> Fix version: 2.7.0 >> >> >> >> Description: >> >> Apache Hadoop contains a flaw in a servlet on the DataNode related to the HDFS namespace browsing functionality. The issue is triggered as a query parameter is not properly validated when handling NameNode information. This may allow a remote attacker to have an unspecified impact. >> >> >> >> References: >> >> https://cve.mitre.org/cgi-bin/cvename.cgi?name=2017-3162 >> >> https://seclists.org/oss-sec/2017/q2/126 >> >> >> >> 12.Package: hadoop-2.6.5 >> >> Severity: HIGH >> >> >> >> Description: >> >> In Apache Hadoop 3.1.0 to 3.1.1, 3.0.0-alpha1 to 3.0.3, 2.9.0 to 2.9.1, and 2.0.0-alpha to 2.8.4, the user/group information can be corrupted across storing in fsimage and reading back from fsimage. >> >> >> >> References: >> >> https://nvd.nist.gov/vuln/detail/CVE-2018-11768 >> >> >> >> Paths: >> >> Hadoop paths >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-co >> mmon >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-cl >> ient >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-ap >> i >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-mapredu >> ce-client-core >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-hdfs >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-common >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-auth >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-annotat >> ions >> >> >> >> 13.Package: protobuf- 2.5.0, 2.6.1 >> >> Severity: HIGH >> >> >> >> Description: >> >> protobuf allows remote authenticated attackers to cause a heap-based buffer overflow. >> >> >> >> Path: >> >> /opt/flink/lib/flink-dist_2.11-1.10.1.jar:protobuf-java >> >> /opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:protobuf-java >> >> >> >> References: >> >> https://nvd.nist.gov/vuln/detail/CVE-2015-5237 >> >> >> >> Please let us know your comments on these issues and fix plans. >> >> >> >> Regards, >> >> Suchithra >> >> |
Free forum by Nabble | Edit this page |