commit
705a25c122
26 changed files with 1988 additions and 0 deletions
-
33.gitignore
-
1README.md
-
308mvnw
-
205mvnw.cmd
-
119pom.xml
-
13src/main/java/com/bfd/crawl/kafkahandler/KafkaHandlerApplication.java
-
61src/main/java/com/bfd/crawl/kafkahandler/bean/ResponsePo.java
-
63src/main/java/com/bfd/crawl/kafkahandler/config/AsyncThreadConfiguration.java
-
60src/main/java/com/bfd/crawl/kafkahandler/config/Constant.java
-
25src/main/java/com/bfd/crawl/kafkahandler/config/ZookeeperConfig.java
-
53src/main/java/com/bfd/crawl/kafkahandler/controller/HandlerDataController.java
-
32src/main/java/com/bfd/crawl/kafkahandler/enums/ResponseCode.java
-
160src/main/java/com/bfd/crawl/kafkahandler/service/ConsumerHandler.java
-
73src/main/java/com/bfd/crawl/kafkahandler/service/ConsumerHandlerService.java
-
93src/main/java/com/bfd/crawl/kafkahandler/service/ProducterHandlerService.java
-
102src/main/java/com/bfd/crawl/kafkahandler/service/SchHandlerService.java
-
61src/main/java/com/bfd/crawl/kafkahandler/service/SendService.java
-
60src/main/java/com/bfd/crawl/kafkahandler/service/StartServcie.java
-
66src/main/java/com/bfd/crawl/kafkahandler/service/TimeOutHandlerService.java
-
66src/main/java/com/bfd/crawl/kafkahandler/service/ZookeeperNodeMonitor.java
-
65src/main/java/com/bfd/crawl/kafkahandler/util/DataUtil.java
-
16src/main/java/com/bfd/crawl/kafkahandler/util/QueueUtil.java
-
147src/main/java/com/bfd/crawl/kafkahandler/util/StringUtil.java
-
57src/main/resources/application.yml
-
36src/main/resources/logback-spring.xml
-
13src/test/java/com/bfd/crawl/kafkahandler/KafkaHandlerApplicationTests.java
@ -0,0 +1,33 @@ |
|||||
|
HELP.md |
||||
|
target/ |
||||
|
!.mvn/wrapper/maven-wrapper.jar |
||||
|
!**/src/main/**/target/ |
||||
|
!**/src/test/**/target/ |
||||
|
|
||||
|
### STS ### |
||||
|
.apt_generated |
||||
|
.classpath |
||||
|
.factorypath |
||||
|
.project |
||||
|
.settings |
||||
|
.springBeans |
||||
|
.sts4-cache |
||||
|
|
||||
|
### IntelliJ IDEA ### |
||||
|
.idea |
||||
|
*.iws |
||||
|
*.iml |
||||
|
*.ipr |
||||
|
|
||||
|
### NetBeans ### |
||||
|
/nbproject/private/ |
||||
|
/nbbuild/ |
||||
|
/dist/ |
||||
|
/nbdist/ |
||||
|
/.nb-gradle/ |
||||
|
build/ |
||||
|
!**/src/main/**/build/ |
||||
|
!**/src/test/**/build/ |
||||
|
|
||||
|
### VS Code ### |
||||
|
.vscode/ |
@ -0,0 +1 @@ |
|||||
|
kafka读取、写入应用 |
@ -0,0 +1,308 @@ |
|||||
|
#!/bin/sh |
||||
|
# ---------------------------------------------------------------------------- |
||||
|
# Licensed to the Apache Software Foundation (ASF) under one |
||||
|
# or more contributor license agreements. See the NOTICE file |
||||
|
# distributed with this work for additional information |
||||
|
# regarding copyright ownership. The ASF licenses this file |
||||
|
# to you under the Apache License, Version 2.0 (the |
||||
|
# "License"); you may not use this file except in compliance |
||||
|
# with the License. You may obtain a copy of the License at |
||||
|
# |
||||
|
# https://www.apache.org/licenses/LICENSE-2.0 |
||||
|
# |
||||
|
# Unless required by applicable law or agreed to in writing, |
||||
|
# software distributed under the License is distributed on an |
||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
||||
|
# KIND, either express or implied. See the License for the |
||||
|
# specific language governing permissions and limitations |
||||
|
# under the License. |
||||
|
# ---------------------------------------------------------------------------- |
||||
|
|
||||
|
# ---------------------------------------------------------------------------- |
||||
|
# Apache Maven Wrapper startup batch script, version 3.2.0 |
||||
|
# |
||||
|
# Required ENV vars: |
||||
|
# ------------------ |
||||
|
# JAVA_HOME - location of a JDK home dir |
||||
|
# |
||||
|
# Optional ENV vars |
||||
|
# ----------------- |
||||
|
# MAVEN_OPTS - parameters passed to the Java VM when running Maven |
||||
|
# e.g. to debug Maven itself, use |
||||
|
# set MAVEN_OPTS=-Xdebug -Xrunjdwp:transport=dt_socket,server=y,suspend=y,address=8000 |
||||
|
# MAVEN_SKIP_RC - flag to disable loading of mavenrc files |
||||
|
# ---------------------------------------------------------------------------- |
||||
|
|
||||
|
if [ -z "$MAVEN_SKIP_RC" ] ; then |
||||
|
|
||||
|
if [ -f /usr/local/etc/mavenrc ] ; then |
||||
|
. /usr/local/etc/mavenrc |
||||
|
fi |
||||
|
|
||||
|
if [ -f /etc/mavenrc ] ; then |
||||
|
. /etc/mavenrc |
||||
|
fi |
||||
|
|
||||
|
if [ -f "$HOME/.mavenrc" ] ; then |
||||
|
. "$HOME/.mavenrc" |
||||
|
fi |
||||
|
|
||||
|
fi |
||||
|
|
||||
|
# OS specific support. $var _must_ be set to either true or false. |
||||
|
cygwin=false; |
||||
|
darwin=false; |
||||
|
mingw=false |
||||
|
case "$(uname)" in |
||||
|
CYGWIN*) cygwin=true ;; |
||||
|
MINGW*) mingw=true;; |
||||
|
Darwin*) darwin=true |
||||
|
# Use /usr/libexec/java_home if available, otherwise fall back to /Library/Java/Home |
||||
|
# See https://developer.apple.com/library/mac/qa/qa1170/_index.html |
||||
|
if [ -z "$JAVA_HOME" ]; then |
||||
|
if [ -x "/usr/libexec/java_home" ]; then |
||||
|
JAVA_HOME="$(/usr/libexec/java_home)"; export JAVA_HOME |
||||
|
else |
||||
|
JAVA_HOME="/Library/Java/Home"; export JAVA_HOME |
||||
|
fi |
||||
|
fi |
||||
|
;; |
||||
|
esac |
||||
|
|
||||
|
if [ -z "$JAVA_HOME" ] ; then |
||||
|
if [ -r /etc/gentoo-release ] ; then |
||||
|
JAVA_HOME=$(java-config --jre-home) |
||||
|
fi |
||||
|
fi |
||||
|
|
||||
|
# For Cygwin, ensure paths are in UNIX format before anything is touched |
||||
|
if $cygwin ; then |
||||
|
[ -n "$JAVA_HOME" ] && |
||||
|
JAVA_HOME=$(cygpath --unix "$JAVA_HOME") |
||||
|
[ -n "$CLASSPATH" ] && |
||||
|
CLASSPATH=$(cygpath --path --unix "$CLASSPATH") |
||||
|
fi |
||||
|
|
||||
|
# For Mingw, ensure paths are in UNIX format before anything is touched |
||||
|
if $mingw ; then |
||||
|
[ -n "$JAVA_HOME" ] && [ -d "$JAVA_HOME" ] && |
||||
|
JAVA_HOME="$(cd "$JAVA_HOME" || (echo "cannot cd into $JAVA_HOME."; exit 1); pwd)" |
||||
|
fi |
||||
|
|
||||
|
if [ -z "$JAVA_HOME" ]; then |
||||
|
javaExecutable="$(which javac)" |
||||
|
if [ -n "$javaExecutable" ] && ! [ "$(expr "\"$javaExecutable\"" : '\([^ ]*\)')" = "no" ]; then |
||||
|
# readlink(1) is not available as standard on Solaris 10. |
||||
|
readLink=$(which readlink) |
||||
|
if [ ! "$(expr "$readLink" : '\([^ ]*\)')" = "no" ]; then |
||||
|
if $darwin ; then |
||||
|
javaHome="$(dirname "\"$javaExecutable\"")" |
||||
|
javaExecutable="$(cd "\"$javaHome\"" && pwd -P)/javac" |
||||
|
else |
||||
|
javaExecutable="$(readlink -f "\"$javaExecutable\"")" |
||||
|
fi |
||||
|
javaHome="$(dirname "\"$javaExecutable\"")" |
||||
|
javaHome=$(expr "$javaHome" : '\(.*\)/bin') |
||||
|
JAVA_HOME="$javaHome" |
||||
|
export JAVA_HOME |
||||
|
fi |
||||
|
fi |
||||
|
fi |
||||
|
|
||||
|
if [ -z "$JAVACMD" ] ; then |
||||
|
if [ -n "$JAVA_HOME" ] ; then |
||||
|
if [ -x "$JAVA_HOME/jre/sh/java" ] ; then |
||||
|
# IBM's JDK on AIX uses strange locations for the executables |
||||
|
JAVACMD="$JAVA_HOME/jre/sh/java" |
||||
|
else |
||||
|
JAVACMD="$JAVA_HOME/bin/java" |
||||
|
fi |
||||
|
else |
||||
|
JAVACMD="$(\unset -f command 2>/dev/null; \command -v java)" |
||||
|
fi |
||||
|
fi |
||||
|
|
||||
|
if [ ! -x "$JAVACMD" ] ; then |
||||
|
echo "Error: JAVA_HOME is not defined correctly." >&2 |
||||
|
echo " We cannot execute $JAVACMD" >&2 |
||||
|
exit 1 |
||||
|
fi |
||||
|
|
||||
|
if [ -z "$JAVA_HOME" ] ; then |
||||
|
echo "Warning: JAVA_HOME environment variable is not set." |
||||
|
fi |
||||
|
|
||||
|
# traverses directory structure from process work directory to filesystem root |
||||
|
# first directory with .mvn subdirectory is considered project base directory |
||||
|
find_maven_basedir() { |
||||
|
if [ -z "$1" ] |
||||
|
then |
||||
|
echo "Path not specified to find_maven_basedir" |
||||
|
return 1 |
||||
|
fi |
||||
|
|
||||
|
basedir="$1" |
||||
|
wdir="$1" |
||||
|
while [ "$wdir" != '/' ] ; do |
||||
|
if [ -d "$wdir"/.mvn ] ; then |
||||
|
basedir=$wdir |
||||
|
break |
||||
|
fi |
||||
|
# workaround for JBEAP-8937 (on Solaris 10/Sparc) |
||||
|
if [ -d "${wdir}" ]; then |
||||
|
wdir=$(cd "$wdir/.." || exit 1; pwd) |
||||
|
fi |
||||
|
# end of workaround |
||||
|
done |
||||
|
printf '%s' "$(cd "$basedir" || exit 1; pwd)" |
||||
|
} |
||||
|
|
||||
|
# concatenates all lines of a file |
||||
|
concat_lines() { |
||||
|
if [ -f "$1" ]; then |
||||
|
# Remove \r in case we run on Windows within Git Bash |
||||
|
# and check out the repository with auto CRLF management |
||||
|
# enabled. Otherwise, we may read lines that are delimited with |
||||
|
# \r\n and produce $'-Xarg\r' rather than -Xarg due to word |
||||
|
# splitting rules. |
||||
|
tr -s '\r\n' ' ' < "$1" |
||||
|
fi |
||||
|
} |
||||
|
|
||||
|
log() { |
||||
|
if [ "$MVNW_VERBOSE" = true ]; then |
||||
|
printf '%s\n' "$1" |
||||
|
fi |
||||
|
} |
||||
|
|
||||
|
BASE_DIR=$(find_maven_basedir "$(dirname "$0")") |
||||
|
if [ -z "$BASE_DIR" ]; then |
||||
|
exit 1; |
||||
|
fi |
||||
|
|
||||
|
MAVEN_PROJECTBASEDIR=${MAVEN_BASEDIR:-"$BASE_DIR"}; export MAVEN_PROJECTBASEDIR |
||||
|
log "$MAVEN_PROJECTBASEDIR" |
||||
|
|
||||
|
########################################################################################## |
||||
|
# Extension to allow automatically downloading the maven-wrapper.jar from Maven-central |
||||
|
# This allows using the maven wrapper in projects that prohibit checking in binary data. |
||||
|
########################################################################################## |
||||
|
wrapperJarPath="$MAVEN_PROJECTBASEDIR/.mvn/wrapper/maven-wrapper.jar" |
||||
|
if [ -r "$wrapperJarPath" ]; then |
||||
|
log "Found $wrapperJarPath" |
||||
|
else |
||||
|
log "Couldn't find $wrapperJarPath, downloading it ..." |
||||
|
|
||||
|
if [ -n "$MVNW_REPOURL" ]; then |
||||
|
wrapperUrl="$MVNW_REPOURL/org/apache/maven/wrapper/maven-wrapper/3.2.0/maven-wrapper-3.2.0.jar" |
||||
|
else |
||||
|
wrapperUrl="https://repo.maven.apache.org/maven2/org/apache/maven/wrapper/maven-wrapper/3.2.0/maven-wrapper-3.2.0.jar" |
||||
|
fi |
||||
|
while IFS="=" read -r key value; do |
||||
|
# Remove '\r' from value to allow usage on windows as IFS does not consider '\r' as a separator ( considers space, tab, new line ('\n'), and custom '=' ) |
||||
|
safeValue=$(echo "$value" | tr -d '\r') |
||||
|
case "$key" in (wrapperUrl) wrapperUrl="$safeValue"; break ;; |
||||
|
esac |
||||
|
done < "$MAVEN_PROJECTBASEDIR/.mvn/wrapper/maven-wrapper.properties" |
||||
|
log "Downloading from: $wrapperUrl" |
||||
|
|
||||
|
if $cygwin; then |
||||
|
wrapperJarPath=$(cygpath --path --windows "$wrapperJarPath") |
||||
|
fi |
||||
|
|
||||
|
if command -v wget > /dev/null; then |
||||
|
log "Found wget ... using wget" |
||||
|
[ "$MVNW_VERBOSE" = true ] && QUIET="" || QUIET="--quiet" |
||||
|
if [ -z "$MVNW_USERNAME" ] || [ -z "$MVNW_PASSWORD" ]; then |
||||
|
wget $QUIET "$wrapperUrl" -O "$wrapperJarPath" || rm -f "$wrapperJarPath" |
||||
|
else |
||||
|
wget $QUIET --http-user="$MVNW_USERNAME" --http-password="$MVNW_PASSWORD" "$wrapperUrl" -O "$wrapperJarPath" || rm -f "$wrapperJarPath" |
||||
|
fi |
||||
|
elif command -v curl > /dev/null; then |
||||
|
log "Found curl ... using curl" |
||||
|
[ "$MVNW_VERBOSE" = true ] && QUIET="" || QUIET="--silent" |
||||
|
if [ -z "$MVNW_USERNAME" ] || [ -z "$MVNW_PASSWORD" ]; then |
||||
|
curl $QUIET -o "$wrapperJarPath" "$wrapperUrl" -f -L || rm -f "$wrapperJarPath" |
||||
|
else |
||||
|
curl $QUIET --user "$MVNW_USERNAME:$MVNW_PASSWORD" -o "$wrapperJarPath" "$wrapperUrl" -f -L || rm -f "$wrapperJarPath" |
||||
|
fi |
||||
|
else |
||||
|
log "Falling back to using Java to download" |
||||
|
javaSource="$MAVEN_PROJECTBASEDIR/.mvn/wrapper/MavenWrapperDownloader.java" |
||||
|
javaClass="$MAVEN_PROJECTBASEDIR/.mvn/wrapper/MavenWrapperDownloader.class" |
||||
|
# For Cygwin, switch paths to Windows format before running javac |
||||
|
if $cygwin; then |
||||
|
javaSource=$(cygpath --path --windows "$javaSource") |
||||
|
javaClass=$(cygpath --path --windows "$javaClass") |
||||
|
fi |
||||
|
if [ -e "$javaSource" ]; then |
||||
|
if [ ! -e "$javaClass" ]; then |
||||
|
log " - Compiling MavenWrapperDownloader.java ..." |
||||
|
("$JAVA_HOME/bin/javac" "$javaSource") |
||||
|
fi |
||||
|
if [ -e "$javaClass" ]; then |
||||
|
log " - Running MavenWrapperDownloader.java ..." |
||||
|
("$JAVA_HOME/bin/java" -cp .mvn/wrapper MavenWrapperDownloader "$wrapperUrl" "$wrapperJarPath") || rm -f "$wrapperJarPath" |
||||
|
fi |
||||
|
fi |
||||
|
fi |
||||
|
fi |
||||
|
########################################################################################## |
||||
|
# End of extension |
||||
|
########################################################################################## |
||||
|
|
||||
|
# If specified, validate the SHA-256 sum of the Maven wrapper jar file |
||||
|
wrapperSha256Sum="" |
||||
|
while IFS="=" read -r key value; do |
||||
|
case "$key" in (wrapperSha256Sum) wrapperSha256Sum=$value; break ;; |
||||
|
esac |
||||
|
done < "$MAVEN_PROJECTBASEDIR/.mvn/wrapper/maven-wrapper.properties" |
||||
|
if [ -n "$wrapperSha256Sum" ]; then |
||||
|
wrapperSha256Result=false |
||||
|
if command -v sha256sum > /dev/null; then |
||||
|
if echo "$wrapperSha256Sum $wrapperJarPath" | sha256sum -c > /dev/null 2>&1; then |
||||
|
wrapperSha256Result=true |
||||
|
fi |
||||
|
elif command -v shasum > /dev/null; then |
||||
|
if echo "$wrapperSha256Sum $wrapperJarPath" | shasum -a 256 -c > /dev/null 2>&1; then |
||||
|
wrapperSha256Result=true |
||||
|
fi |
||||
|
else |
||||
|
echo "Checksum validation was requested but neither 'sha256sum' or 'shasum' are available." |
||||
|
echo "Please install either command, or disable validation by removing 'wrapperSha256Sum' from your maven-wrapper.properties." |
||||
|
exit 1 |
||||
|
fi |
||||
|
if [ $wrapperSha256Result = false ]; then |
||||
|
echo "Error: Failed to validate Maven wrapper SHA-256, your Maven wrapper might be compromised." >&2 |
||||
|
echo "Investigate or delete $wrapperJarPath to attempt a clean download." >&2 |
||||
|
echo "If you updated your Maven version, you need to update the specified wrapperSha256Sum property." >&2 |
||||
|
exit 1 |
||||
|
fi |
||||
|
fi |
||||
|
|
||||
|
MAVEN_OPTS="$(concat_lines "$MAVEN_PROJECTBASEDIR/.mvn/jvm.config") $MAVEN_OPTS" |
||||
|
|
||||
|
# For Cygwin, switch paths to Windows format before running java |
||||
|
if $cygwin; then |
||||
|
[ -n "$JAVA_HOME" ] && |
||||
|
JAVA_HOME=$(cygpath --path --windows "$JAVA_HOME") |
||||
|
[ -n "$CLASSPATH" ] && |
||||
|
CLASSPATH=$(cygpath --path --windows "$CLASSPATH") |
||||
|
[ -n "$MAVEN_PROJECTBASEDIR" ] && |
||||
|
MAVEN_PROJECTBASEDIR=$(cygpath --path --windows "$MAVEN_PROJECTBASEDIR") |
||||
|
fi |
||||
|
|
||||
|
# Provide a "standardized" way to retrieve the CLI args that will |
||||
|
# work with both Windows and non-Windows executions. |
||||
|
MAVEN_CMD_LINE_ARGS="$MAVEN_CONFIG $*" |
||||
|
export MAVEN_CMD_LINE_ARGS |
||||
|
|
||||
|
WRAPPER_LAUNCHER=org.apache.maven.wrapper.MavenWrapperMain |
||||
|
|
||||
|
# shellcheck disable=SC2086 # safe args |
||||
|
exec "$JAVACMD" \ |
||||
|
$MAVEN_OPTS \ |
||||
|
$MAVEN_DEBUG_OPTS \ |
||||
|
-classpath "$MAVEN_PROJECTBASEDIR/.mvn/wrapper/maven-wrapper.jar" \ |
||||
|
"-Dmaven.multiModuleProjectDirectory=${MAVEN_PROJECTBASEDIR}" \ |
||||
|
${WRAPPER_LAUNCHER} $MAVEN_CONFIG "$@" |
@ -0,0 +1,205 @@ |
|||||
|
@REM ---------------------------------------------------------------------------- |
||||
|
@REM Licensed to the Apache Software Foundation (ASF) under one |
||||
|
@REM or more contributor license agreements. See the NOTICE file |
||||
|
@REM distributed with this work for additional information |
||||
|
@REM regarding copyright ownership. The ASF licenses this file |
||||
|
@REM to you under the Apache License, Version 2.0 (the |
||||
|
@REM "License"); you may not use this file except in compliance |
||||
|
@REM with the License. You may obtain a copy of the License at |
||||
|
@REM |
||||
|
@REM https://www.apache.org/licenses/LICENSE-2.0 |
||||
|
@REM |
||||
|
@REM Unless required by applicable law or agreed to in writing, |
||||
|
@REM software distributed under the License is distributed on an |
||||
|
@REM "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
||||
|
@REM KIND, either express or implied. See the License for the |
||||
|
@REM specific language governing permissions and limitations |
||||
|
@REM under the License. |
||||
|
@REM ---------------------------------------------------------------------------- |
||||
|
|
||||
|
@REM ---------------------------------------------------------------------------- |
||||
|
@REM Apache Maven Wrapper startup batch script, version 3.2.0 |
||||
|
@REM |
||||
|
@REM Required ENV vars: |
||||
|
@REM JAVA_HOME - location of a JDK home dir |
||||
|
@REM |
||||
|
@REM Optional ENV vars |
||||
|
@REM MAVEN_BATCH_ECHO - set to 'on' to enable the echoing of the batch commands |
||||
|
@REM MAVEN_BATCH_PAUSE - set to 'on' to wait for a keystroke before ending |
||||
|
@REM MAVEN_OPTS - parameters passed to the Java VM when running Maven |
||||
|
@REM e.g. to debug Maven itself, use |
||||
|
@REM set MAVEN_OPTS=-Xdebug -Xrunjdwp:transport=dt_socket,server=y,suspend=y,address=8000 |
||||
|
@REM MAVEN_SKIP_RC - flag to disable loading of mavenrc files |
||||
|
@REM ---------------------------------------------------------------------------- |
||||
|
|
||||
|
@REM Begin all REM lines with '@' in case MAVEN_BATCH_ECHO is 'on' |
||||
|
@echo off |
||||
|
@REM set title of command window |
||||
|
title %0 |
||||
|
@REM enable echoing by setting MAVEN_BATCH_ECHO to 'on' |
||||
|
@if "%MAVEN_BATCH_ECHO%" == "on" echo %MAVEN_BATCH_ECHO% |
||||
|
|
||||
|
@REM set %HOME% to equivalent of $HOME |
||||
|
if "%HOME%" == "" (set "HOME=%HOMEDRIVE%%HOMEPATH%") |
||||
|
|
||||
|
@REM Execute a user defined script before this one |
||||
|
if not "%MAVEN_SKIP_RC%" == "" goto skipRcPre |
||||
|
@REM check for pre script, once with legacy .bat ending and once with .cmd ending |
||||
|
if exist "%USERPROFILE%\mavenrc_pre.bat" call "%USERPROFILE%\mavenrc_pre.bat" %* |
||||
|
if exist "%USERPROFILE%\mavenrc_pre.cmd" call "%USERPROFILE%\mavenrc_pre.cmd" %* |
||||
|
:skipRcPre |
||||
|
|
||||
|
@setlocal |
||||
|
|
||||
|
set ERROR_CODE=0 |
||||
|
|
||||
|
@REM To isolate internal variables from possible post scripts, we use another setlocal |
||||
|
@setlocal |
||||
|
|
||||
|
@REM ==== START VALIDATION ==== |
||||
|
if not "%JAVA_HOME%" == "" goto OkJHome |
||||
|
|
||||
|
echo. |
||||
|
echo Error: JAVA_HOME not found in your environment. >&2 |
||||
|
echo Please set the JAVA_HOME variable in your environment to match the >&2 |
||||
|
echo location of your Java installation. >&2 |
||||
|
echo. |
||||
|
goto error |
||||
|
|
||||
|
:OkJHome |
||||
|
if exist "%JAVA_HOME%\bin\java.exe" goto init |
||||
|
|
||||
|
echo. |
||||
|
echo Error: JAVA_HOME is set to an invalid directory. >&2 |
||||
|
echo JAVA_HOME = "%JAVA_HOME%" >&2 |
||||
|
echo Please set the JAVA_HOME variable in your environment to match the >&2 |
||||
|
echo location of your Java installation. >&2 |
||||
|
echo. |
||||
|
goto error |
||||
|
|
||||
|
@REM ==== END VALIDATION ==== |
||||
|
|
||||
|
:init |
||||
|
|
||||
|
@REM Find the project base dir, i.e. the directory that contains the folder ".mvn". |
||||
|
@REM Fallback to current working directory if not found. |
||||
|
|
||||
|
set MAVEN_PROJECTBASEDIR=%MAVEN_BASEDIR% |
||||
|
IF NOT "%MAVEN_PROJECTBASEDIR%"=="" goto endDetectBaseDir |
||||
|
|
||||
|
set EXEC_DIR=%CD% |
||||
|
set WDIR=%EXEC_DIR% |
||||
|
:findBaseDir |
||||
|
IF EXIST "%WDIR%"\.mvn goto baseDirFound |
||||
|
cd .. |
||||
|
IF "%WDIR%"=="%CD%" goto baseDirNotFound |
||||
|
set WDIR=%CD% |
||||
|
goto findBaseDir |
||||
|
|
||||
|
:baseDirFound |
||||
|
set MAVEN_PROJECTBASEDIR=%WDIR% |
||||
|
cd "%EXEC_DIR%" |
||||
|
goto endDetectBaseDir |
||||
|
|
||||
|
:baseDirNotFound |
||||
|
set MAVEN_PROJECTBASEDIR=%EXEC_DIR% |
||||
|
cd "%EXEC_DIR%" |
||||
|
|
||||
|
:endDetectBaseDir |
||||
|
|
||||
|
IF NOT EXIST "%MAVEN_PROJECTBASEDIR%\.mvn\jvm.config" goto endReadAdditionalConfig |
||||
|
|
||||
|
@setlocal EnableExtensions EnableDelayedExpansion |
||||
|
for /F "usebackq delims=" %%a in ("%MAVEN_PROJECTBASEDIR%\.mvn\jvm.config") do set JVM_CONFIG_MAVEN_PROPS=!JVM_CONFIG_MAVEN_PROPS! %%a |
||||
|
@endlocal & set JVM_CONFIG_MAVEN_PROPS=%JVM_CONFIG_MAVEN_PROPS% |
||||
|
|
||||
|
:endReadAdditionalConfig |
||||
|
|
||||
|
SET MAVEN_JAVA_EXE="%JAVA_HOME%\bin\java.exe" |
||||
|
set WRAPPER_JAR="%MAVEN_PROJECTBASEDIR%\.mvn\wrapper\maven-wrapper.jar" |
||||
|
set WRAPPER_LAUNCHER=org.apache.maven.wrapper.MavenWrapperMain |
||||
|
|
||||
|
set WRAPPER_URL="https://repo.maven.apache.org/maven2/org/apache/maven/wrapper/maven-wrapper/3.2.0/maven-wrapper-3.2.0.jar" |
||||
|
|
||||
|
FOR /F "usebackq tokens=1,2 delims==" %%A IN ("%MAVEN_PROJECTBASEDIR%\.mvn\wrapper\maven-wrapper.properties") DO ( |
||||
|
IF "%%A"=="wrapperUrl" SET WRAPPER_URL=%%B |
||||
|
) |
||||
|
|
||||
|
@REM Extension to allow automatically downloading the maven-wrapper.jar from Maven-central |
||||
|
@REM This allows using the maven wrapper in projects that prohibit checking in binary data. |
||||
|
if exist %WRAPPER_JAR% ( |
||||
|
if "%MVNW_VERBOSE%" == "true" ( |
||||
|
echo Found %WRAPPER_JAR% |
||||
|
) |
||||
|
) else ( |
||||
|
if not "%MVNW_REPOURL%" == "" ( |
||||
|
SET WRAPPER_URL="%MVNW_REPOURL%/org/apache/maven/wrapper/maven-wrapper/3.2.0/maven-wrapper-3.2.0.jar" |
||||
|
) |
||||
|
if "%MVNW_VERBOSE%" == "true" ( |
||||
|
echo Couldn't find %WRAPPER_JAR%, downloading it ... |
||||
|
echo Downloading from: %WRAPPER_URL% |
||||
|
) |
||||
|
|
||||
|
powershell -Command "&{"^ |
||||
|
"$webclient = new-object System.Net.WebClient;"^ |
||||
|
"if (-not ([string]::IsNullOrEmpty('%MVNW_USERNAME%') -and [string]::IsNullOrEmpty('%MVNW_PASSWORD%'))) {"^ |
||||
|
"$webclient.Credentials = new-object System.Net.NetworkCredential('%MVNW_USERNAME%', '%MVNW_PASSWORD%');"^ |
||||
|
"}"^ |
||||
|
"[Net.ServicePointManager]::SecurityProtocol = [Net.SecurityProtocolType]::Tls12; $webclient.DownloadFile('%WRAPPER_URL%', '%WRAPPER_JAR%')"^ |
||||
|
"}" |
||||
|
if "%MVNW_VERBOSE%" == "true" ( |
||||
|
echo Finished downloading %WRAPPER_JAR% |
||||
|
) |
||||
|
) |
||||
|
@REM End of extension |
||||
|
|
||||
|
@REM If specified, validate the SHA-256 sum of the Maven wrapper jar file |
||||
|
SET WRAPPER_SHA_256_SUM="" |
||||
|
FOR /F "usebackq tokens=1,2 delims==" %%A IN ("%MAVEN_PROJECTBASEDIR%\.mvn\wrapper\maven-wrapper.properties") DO ( |
||||
|
IF "%%A"=="wrapperSha256Sum" SET WRAPPER_SHA_256_SUM=%%B |
||||
|
) |
||||
|
IF NOT %WRAPPER_SHA_256_SUM%=="" ( |
||||
|
powershell -Command "&{"^ |
||||
|
"$hash = (Get-FileHash \"%WRAPPER_JAR%\" -Algorithm SHA256).Hash.ToLower();"^ |
||||
|
"If('%WRAPPER_SHA_256_SUM%' -ne $hash){"^ |
||||
|
" Write-Output 'Error: Failed to validate Maven wrapper SHA-256, your Maven wrapper might be compromised.';"^ |
||||
|
" Write-Output 'Investigate or delete %WRAPPER_JAR% to attempt a clean download.';"^ |
||||
|
" Write-Output 'If you updated your Maven version, you need to update the specified wrapperSha256Sum property.';"^ |
||||
|
" exit 1;"^ |
||||
|
"}"^ |
||||
|
"}" |
||||
|
if ERRORLEVEL 1 goto error |
||||
|
) |
||||
|
|
||||
|
@REM Provide a "standardized" way to retrieve the CLI args that will |
||||
|
@REM work with both Windows and non-Windows executions. |
||||
|
set MAVEN_CMD_LINE_ARGS=%* |
||||
|
|
||||
|
%MAVEN_JAVA_EXE% ^ |
||||
|
%JVM_CONFIG_MAVEN_PROPS% ^ |
||||
|
%MAVEN_OPTS% ^ |
||||
|
%MAVEN_DEBUG_OPTS% ^ |
||||
|
-classpath %WRAPPER_JAR% ^ |
||||
|
"-Dmaven.multiModuleProjectDirectory=%MAVEN_PROJECTBASEDIR%" ^ |
||||
|
%WRAPPER_LAUNCHER% %MAVEN_CONFIG% %* |
||||
|
if ERRORLEVEL 1 goto error |
||||
|
goto end |
||||
|
|
||||
|
:error |
||||
|
set ERROR_CODE=1 |
||||
|
|
||||
|
:end |
||||
|
@endlocal & set ERROR_CODE=%ERROR_CODE% |
||||
|
|
||||
|
if not "%MAVEN_SKIP_RC%"=="" goto skipRcPost |
||||
|
@REM check for post script, once with legacy .bat ending and once with .cmd ending |
||||
|
if exist "%USERPROFILE%\mavenrc_post.bat" call "%USERPROFILE%\mavenrc_post.bat" |
||||
|
if exist "%USERPROFILE%\mavenrc_post.cmd" call "%USERPROFILE%\mavenrc_post.cmd" |
||||
|
:skipRcPost |
||||
|
|
||||
|
@REM pause the script if MAVEN_BATCH_PAUSE is set to 'on' |
||||
|
if "%MAVEN_BATCH_PAUSE%"=="on" pause |
||||
|
|
||||
|
if "%MAVEN_TERMINATE_CMD%"=="on" exit %ERROR_CODE% |
||||
|
|
||||
|
cmd /C exit /B %ERROR_CODE% |
@ -0,0 +1,119 @@ |
|||||
|
<?xml version="1.0" encoding="UTF-8"?> |
||||
|
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" |
||||
|
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 https://maven.apache.org/xsd/maven-4.0.0.xsd"> |
||||
|
<modelVersion>4.0.0</modelVersion> |
||||
|
<parent> |
||||
|
<groupId>org.springframework.boot</groupId> |
||||
|
<artifactId>spring-boot-starter-parent</artifactId> |
||||
|
<version>2.2.4.RELEASE</version> |
||||
|
<relativePath/> <!-- lookup parent from repository --> |
||||
|
</parent> |
||||
|
<groupId>com.bfd.crawl</groupId> |
||||
|
<artifactId>kafkaHandler</artifactId> |
||||
|
<version>0.0.1-SNAPSHOT</version> |
||||
|
<name>kafkaHandler</name> |
||||
|
<description>kafkaHandler</description> |
||||
|
<properties> |
||||
|
<java.version>8</java.version> |
||||
|
</properties> |
||||
|
<dependencies> |
||||
|
<dependency> |
||||
|
<groupId>de.codecentric</groupId> |
||||
|
<artifactId>spring-boot-admin-client</artifactId> |
||||
|
<version>2.2.4</version> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.springframework.boot</groupId> |
||||
|
<artifactId>spring-boot-starter</artifactId> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.springframework.kafka</groupId> |
||||
|
<artifactId>spring-kafka</artifactId> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.springframework.boot</groupId> |
||||
|
<artifactId>spring-boot-starter-web</artifactId> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>com.google.code.gson</groupId> |
||||
|
<artifactId>gson</artifactId> |
||||
|
<version>2.8.8</version> |
||||
|
</dependency> |
||||
|
<!--JSON--> |
||||
|
<dependency> |
||||
|
<groupId>com.alibaba</groupId> |
||||
|
<artifactId>fastjson</artifactId> |
||||
|
<version>2.0.17</version> |
||||
|
</dependency> |
||||
|
<!--OKHTTP--> |
||||
|
<dependency> |
||||
|
<groupId>com.squareup.okhttp3</groupId> |
||||
|
<artifactId>okhttp</artifactId> |
||||
|
<version>3.9.1</version> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.apache.curator</groupId> |
||||
|
<artifactId>curator-framework</artifactId> |
||||
|
<version>5.2.0</version> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.apache.curator</groupId> |
||||
|
<artifactId>curator-recipes</artifactId> |
||||
|
<version>5.2.0</version> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.springframework.boot</groupId> |
||||
|
<artifactId>spring-boot-devtools</artifactId> |
||||
|
<scope>runtime</scope> |
||||
|
<optional>true</optional> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.projectlombok</groupId> |
||||
|
<artifactId>lombok</artifactId> |
||||
|
<optional>true</optional> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.springframework.boot</groupId> |
||||
|
<artifactId>spring-boot-starter-test</artifactId> |
||||
|
<scope>test</scope> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.apache.kafka</groupId> |
||||
|
<artifactId>kafka-clients</artifactId> |
||||
|
<version>2.3.1</version> |
||||
|
</dependency> |
||||
|
<!--redis--> |
||||
|
<dependency> |
||||
|
<groupId>org.redisson</groupId> |
||||
|
<artifactId>redisson-spring-boot-starter</artifactId> |
||||
|
<version>3.13.6</version> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>org.springframework.boot</groupId> |
||||
|
<artifactId>spring-boot-starter-data-redis</artifactId> |
||||
|
</dependency> |
||||
|
<dependency> |
||||
|
<groupId>com.bfd.util</groupId> |
||||
|
<artifactId>pauseTool</artifactId> |
||||
|
<version>1.0</version> |
||||
|
</dependency> |
||||
|
</dependencies> |
||||
|
|
||||
|
<build> |
||||
|
<plugins> |
||||
|
<plugin> |
||||
|
<groupId>org.springframework.boot</groupId> |
||||
|
<artifactId>spring-boot-maven-plugin</artifactId> |
||||
|
<configuration> |
||||
|
<excludes> |
||||
|
<exclude> |
||||
|
<groupId>org.projectlombok</groupId> |
||||
|
<artifactId>lombok</artifactId> |
||||
|
</exclude> |
||||
|
</excludes> |
||||
|
</configuration> |
||||
|
</plugin> |
||||
|
</plugins> |
||||
|
</build> |
||||
|
|
||||
|
</project> |
@ -0,0 +1,13 @@ |
|||||
|
package com.bfd.crawl.kafkahandler; |
||||
|
|
||||
|
import org.springframework.boot.SpringApplication; |
||||
|
import org.springframework.boot.autoconfigure.SpringBootApplication; |
||||
|
|
||||
|
@SpringBootApplication |
||||
|
public class KafkaHandlerApplication { |
||||
|
|
||||
|
public static void main(String[] args) { |
||||
|
SpringApplication.run(KafkaHandlerApplication.class, args); |
||||
|
} |
||||
|
|
||||
|
} |
@ -0,0 +1,61 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.bean; |
||||
|
|
||||
|
|
||||
|
|
||||
|
|
||||
|
import com.bfd.crawl.kafkahandler.enums.ResponseCode; |
||||
|
import lombok.AllArgsConstructor; |
||||
|
import lombok.Data; |
||||
|
import lombok.NoArgsConstructor; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:ResponsePo |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/4/3 17:23 |
||||
|
*/ |
||||
|
@Data |
||||
|
@NoArgsConstructor |
||||
|
@AllArgsConstructor |
||||
|
public class ResponsePo { |
||||
|
/** |
||||
|
* 响应码 |
||||
|
*/ |
||||
|
private int code; |
||||
|
|
||||
|
/** |
||||
|
* 正常放 返回数据 的JSON串 |
||||
|
*/ |
||||
|
private Object data; |
||||
|
|
||||
|
/** |
||||
|
* 提示消息 |
||||
|
*/ |
||||
|
private String message; |
||||
|
|
||||
|
public static ResponsePo success() { |
||||
|
return setStatus(ResponseCode.SUCCESS.getCode(), ResponseCode.SUCCESS.getMessage()); |
||||
|
} |
||||
|
|
||||
|
public static ResponsePo error() { |
||||
|
return setStatus(ResponseCode.FAILURE.getCode(), ResponseCode.FAILURE.getMessage()); |
||||
|
} |
||||
|
|
||||
|
public static ResponsePo setStatus(int code, String message) { |
||||
|
ResponsePo resultBean = new ResponsePo(); |
||||
|
resultBean.code = code; |
||||
|
resultBean.message = message; |
||||
|
return resultBean; |
||||
|
} |
||||
|
public ResponsePo(int code, String message) { |
||||
|
this.code = code; |
||||
|
this.message = message; |
||||
|
this.data = data; |
||||
|
} |
||||
|
public ResponsePo(ResponseCode responseCode){ |
||||
|
this.code = responseCode.getCode(); |
||||
|
this.message = responseCode.getMessage(); |
||||
|
this.data = data; |
||||
|
} |
||||
|
} |
@ -0,0 +1,63 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.config; |
||||
|
|
||||
|
|
||||
|
import org.springframework.context.annotation.Bean; |
||||
|
import org.springframework.context.annotation.Configuration; |
||||
|
import org.springframework.scheduling.annotation.EnableAsync; |
||||
|
import org.springframework.scheduling.concurrent.ThreadPoolTaskExecutor; |
||||
|
|
||||
|
import java.util.concurrent.Executor; |
||||
|
|
||||
|
/** |
||||
|
* @author jinming |
||||
|
* @version 1.0 |
||||
|
* @className AsyncThreadConfiguration |
||||
|
* @Date 2022/2/17 18:37 |
||||
|
*/ |
||||
|
@Configuration |
||||
|
@EnableAsync |
||||
|
public class AsyncThreadConfiguration { |
||||
|
@Bean |
||||
|
public Executor asyncExecutor() { |
||||
|
ThreadPoolTaskExecutor executor = new ThreadPoolTaskExecutor(); |
||||
|
// 核心线程数 |
||||
|
executor.setCorePoolSize(50); |
||||
|
// 并发线程的数量限制为2 |
||||
|
executor.setMaxPoolSize(50); |
||||
|
// 线程队列 |
||||
|
executor.setQueueCapacity(50); |
||||
|
executor.setThreadNamePrefix("handlerData-"); |
||||
|
executor.initialize(); |
||||
|
executor.setWaitForTasksToCompleteOnShutdown(true); |
||||
|
return executor; |
||||
|
} |
||||
|
@Bean |
||||
|
public Executor producterExecutor() { |
||||
|
ThreadPoolTaskExecutor executor = new ThreadPoolTaskExecutor(); |
||||
|
// 核心线程数 |
||||
|
executor.setCorePoolSize(150); |
||||
|
// 并发线程的数量限制为2 |
||||
|
executor.setMaxPoolSize(150); |
||||
|
// 线程队列 |
||||
|
executor.setQueueCapacity(1000); |
||||
|
executor.setThreadNamePrefix("sendData-"); |
||||
|
executor.initialize(); |
||||
|
executor.setWaitForTasksToCompleteOnShutdown(true); |
||||
|
return executor; |
||||
|
} |
||||
|
|
||||
|
@Bean |
||||
|
public Executor consumerHandlerExecutor() { |
||||
|
ThreadPoolTaskExecutor executor = new ThreadPoolTaskExecutor(); |
||||
|
// 核心线程数 |
||||
|
executor.setCorePoolSize(150); |
||||
|
// 并发线程的数量限制为2 |
||||
|
executor.setMaxPoolSize(150); |
||||
|
// 线程队列 |
||||
|
executor.setQueueCapacity(150); |
||||
|
executor.setThreadNamePrefix("consumerHandler-"); |
||||
|
executor.initialize(); |
||||
|
executor.setWaitForTasksToCompleteOnShutdown(true); |
||||
|
return executor; |
||||
|
} |
||||
|
} |
@ -0,0 +1,60 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.config; |
||||
|
|
||||
|
import org.apache.kafka.clients.consumer.KafkaConsumer; |
||||
|
import org.apache.kafka.clients.producer.KafkaProducer; |
||||
|
|
||||
|
import java.util.HashMap; |
||||
|
import java.util.Map; |
||||
|
import java.util.concurrent.ConcurrentHashMap; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:Constant |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/8/16 15:26 |
||||
|
*/ |
||||
|
public class Constant { |
||||
|
/** |
||||
|
* kafka producter 对象存储 |
||||
|
*/ |
||||
|
public static Map<String, KafkaProducer<String, String>> PRODUCER_MAP = new HashMap<>(32); |
||||
|
|
||||
|
/** |
||||
|
* kafka consumer 对象存储 |
||||
|
*/ |
||||
|
public static Map<String, KafkaConsumer<String, String>> CONSUMER_MAP = new HashMap<>(32); |
||||
|
|
||||
|
/** |
||||
|
* kafka runTimeMap 对象存储 |
||||
|
*/ |
||||
|
public static Map<Integer, Boolean> CONSUMER_RUNTIME_MAP = new HashMap<>(32); |
||||
|
|
||||
|
/** |
||||
|
* kafka consumer 超时问题存储 |
||||
|
*/ |
||||
|
public static ConcurrentHashMap<String, Map> TIME_OUT_MAP = new ConcurrentHashMap<>(32); |
||||
|
|
||||
|
/** |
||||
|
* 生产者的任务类型 |
||||
|
*/ |
||||
|
public final static Integer PRODUCER_TYPE = 1; |
||||
|
|
||||
|
/** |
||||
|
* 空字符串常量 |
||||
|
*/ |
||||
|
public static final String EMPTY = ""; |
||||
|
|
||||
|
/** |
||||
|
* 不需要DataUtil解析的Key |
||||
|
*/ |
||||
|
public static final String NOT_KEY = ":$"; |
||||
|
|
||||
|
|
||||
|
public static final String FORM = "form"; |
||||
|
public static final String FIELD = "field"; |
||||
|
public static final String VALUE = "value"; |
||||
|
public static final String ALL= "*"; |
||||
|
public static final String DATATYPE = "dataType"; |
||||
|
public static final String STRING_TYPE = "String"; |
||||
|
} |
@ -0,0 +1,25 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.config; |
||||
|
|
||||
|
import org.apache.curator.framework.CuratorFramework; |
||||
|
import org.apache.curator.framework.CuratorFrameworkFactory; |
||||
|
import org.apache.curator.retry.ExponentialBackoffRetry; |
||||
|
import org.springframework.beans.factory.annotation.Value; |
||||
|
import org.springframework.context.annotation.Bean; |
||||
|
import org.springframework.context.annotation.Configuration; |
||||
|
|
||||
|
/** |
||||
|
* @author jian.mao |
||||
|
* @date 2024年4月16日 |
||||
|
* @description |
||||
|
*/ |
||||
|
@Configuration |
||||
|
public class ZookeeperConfig { |
||||
|
@Value("${zookeeper.connection-string}") |
||||
|
private String connectionString; |
||||
|
@Bean |
||||
|
public CuratorFramework curatorFramework() { |
||||
|
CuratorFramework curatorFramework = CuratorFrameworkFactory.newClient(connectionString, new ExponentialBackoffRetry(1000, 3)); |
||||
|
curatorFramework.start(); |
||||
|
return curatorFramework; |
||||
|
} |
||||
|
} |
@ -0,0 +1,53 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.controller; |
||||
|
|
||||
|
import com.alibaba.fastjson.JSON; |
||||
|
|
||||
|
import com.bfd.crawl.kafkahandler.bean.ResponsePo; |
||||
|
import com.bfd.crawl.kafkahandler.enums.ResponseCode; |
||||
|
import com.bfd.crawl.kafkahandler.util.QueueUtil; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.springframework.web.bind.annotation.*; |
||||
|
|
||||
|
import java.util.Map; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:HandlerDataController |
||||
|
* @version:1.0 |
||||
|
* @description: 处理接口 |
||||
|
* @Date:2023/7/13 14:25 |
||||
|
*/ |
||||
|
|
||||
|
@RestController |
||||
|
@RequestMapping("/handlerdata") |
||||
|
@Slf4j |
||||
|
public class HandlerDataController { |
||||
|
|
||||
|
@GetMapping("/test") |
||||
|
public String test() { |
||||
|
return "test"; |
||||
|
} |
||||
|
|
||||
|
|
||||
|
|
||||
|
|
||||
|
@PostMapping("/kafka") |
||||
|
public ResponsePo kafkaHandler(@RequestBody String dataJson) { |
||||
|
ResponsePo responsePo = ResponsePo.success(); |
||||
|
try { |
||||
|
Map parse = (Map) JSON.parse(dataJson); |
||||
|
} catch (Exception e) { |
||||
|
log.error("请求格式发生异常" + e.getMessage()); |
||||
|
responsePo.setCode(ResponseCode.FAILURE.getCode()); |
||||
|
responsePo.setMessage(ResponseCode.FAILURE.getMessage()); |
||||
|
return responsePo; |
||||
|
} |
||||
|
log.info("新增任务:" + dataJson); |
||||
|
try { |
||||
|
QueueUtil.taskQueue.put(dataJson); |
||||
|
} catch (InterruptedException e) { |
||||
|
e.printStackTrace(); |
||||
|
} |
||||
|
return responsePo; |
||||
|
} |
||||
|
} |
@ -0,0 +1,32 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.enums; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:ResponseCodeEnum |
||||
|
* @version:1.0 |
||||
|
* @description:响应结果码枚举类 |
||||
|
* @Date:2023/2/28 11:40 |
||||
|
*/ |
||||
|
public enum ResponseCode { |
||||
|
//返回结果码枚举类 |
||||
|
SUCCESS(200, "操作成功"), |
||||
|
FAILURE(400, "参数错误"), |
||||
|
INTERNAL_SERVER_ERROR(500, "服务器内部错误"), |
||||
|
TYPE_NOT_SUPPORT(601,"文件类型不支持"); |
||||
|
|
||||
|
private int code; |
||||
|
private String message; |
||||
|
|
||||
|
ResponseCode(int code, String message) { |
||||
|
this.code = code; |
||||
|
this.message = message; |
||||
|
} |
||||
|
|
||||
|
public int getCode() { |
||||
|
return code; |
||||
|
} |
||||
|
|
||||
|
public String getMessage() { |
||||
|
return message; |
||||
|
} |
||||
|
} |
@ -0,0 +1,160 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.service; |
||||
|
|
||||
|
import com.alibaba.fastjson.JSON; |
||||
|
import com.bfd.crawl.kafkahandler.config.Constant; |
||||
|
import com.bfd.crawl.kafkahandler.util.QueueUtil; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.apache.kafka.clients.consumer.ConsumerRecord; |
||||
|
import org.apache.kafka.clients.consumer.ConsumerRecords; |
||||
|
import org.apache.kafka.clients.consumer.KafkaConsumer; |
||||
|
import org.springframework.data.redis.core.StringRedisTemplate; |
||||
|
import org.springframework.scheduling.annotation.Async; |
||||
|
import org.springframework.stereotype.Service; |
||||
|
|
||||
|
|
||||
|
import javax.annotation.Resource; |
||||
|
import java.util.*; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:ConsumerHandler |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2024/7/9 18:01 |
||||
|
*/ |
||||
|
@Service |
||||
|
@Slf4j |
||||
|
public class ConsumerHandler { |
||||
|
@Resource |
||||
|
private StringRedisTemplate stringRedisTemplate; |
||||
|
|
||||
|
|
||||
|
@Async("consumerHandlerExecutor") |
||||
|
public void creatConsumer(Integer id, String name, Map admin, Map parse, Map output, String objectKey, String groupId) { |
||||
|
try { |
||||
|
int size = (int) admin.get("size"); |
||||
|
String bootstrapServers = (String) admin.get("bootstrapServers"); |
||||
|
|
||||
|
String autoOffsetReset = (String) admin.get("autoOffsetReset"); |
||||
|
String topic = (String) admin.get("topic"); |
||||
|
// 创建Kafka消费者配置 |
||||
|
Properties props = new Properties(); |
||||
|
props.put("bootstrap.servers", bootstrapServers); |
||||
|
props.put("group.id", groupId); |
||||
|
props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); |
||||
|
props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); |
||||
|
props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer"); |
||||
|
props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer"); |
||||
|
props.put("auto.offset.reset", autoOffsetReset); |
||||
|
props.put("request.timeout.ms", "60000"); |
||||
|
props.put("session.timeout.ms", "60000"); |
||||
|
props.put("enable.auto.commit", "false"); |
||||
|
// 创建Kafka消费者 |
||||
|
KafkaConsumer<String, String> consumer = new KafkaConsumer<>(props); |
||||
|
consumer.subscribe(Arrays.asList(topic)); |
||||
|
Constant.CONSUMER_MAP.put(objectKey, consumer); |
||||
|
List<String> filter = (List<String>) admin.get("filter"); |
||||
|
int index = 1; |
||||
|
while (Constant.CONSUMER_RUNTIME_MAP.get(id)) { |
||||
|
boolean isBreak = false; |
||||
|
ConsumerRecords<String, String> msgList = consumer.poll(1000); |
||||
|
for (ConsumerRecord<String, String> record : msgList) { |
||||
|
try { |
||||
|
//取出的任务组装发送至需要流转的Kafka |
||||
|
String value = record.value(); |
||||
|
boolean isContinue = true; |
||||
|
for (String filed : filter) { |
||||
|
if (value.contains(filed)) { |
||||
|
log.info("id:{},kafka满足过滤条件:{}", id, value); |
||||
|
isContinue = false; |
||||
|
} |
||||
|
} |
||||
|
if (isContinue) { |
||||
|
continue; |
||||
|
} |
||||
|
|
||||
|
//fieldType:自定义输出字段: 0 关闭,1-开启,如果开启则拼接form到output里(如果关闭,则取默认的output拼接) |
||||
|
int fieldType = (int) admin.get("fieldType"); |
||||
|
Map result = new HashMap(32); |
||||
|
Map resultsMap = new HashMap(32); |
||||
|
resultsMap.put("result", value); |
||||
|
result.put("results", JSON.toJSONString(resultsMap)); |
||||
|
parse.put("result", result); |
||||
|
if (fieldType != 0) { |
||||
|
resultsMap.remove("result"); |
||||
|
Map valueParse = (Map) JSON.parse(value); |
||||
|
Set set = output.keySet(); |
||||
|
for (Object o : set) { |
||||
|
resultsMap.put(o, valueParse.get(o)); |
||||
|
} |
||||
|
String resultsMapJson = JSON.toJSONString(resultsMap); |
||||
|
result.put("results", resultsMapJson); |
||||
|
parse.put("result", result); |
||||
|
String empty = "{}"; |
||||
|
if (resultsMapJson.equals(empty)) { |
||||
|
continue; |
||||
|
} |
||||
|
} |
||||
|
String message = JSON.toJSONString(parse); |
||||
|
QueueUtil.sendQueue.put(message); |
||||
|
//根据size来判断任务是否要退出,size为-1时代表一直消费,或当timeout到期时关闭consumer导致异常退出 |
||||
|
index++; |
||||
|
if (size != -1 && index == size) { |
||||
|
isBreak = true; |
||||
|
break; |
||||
|
} |
||||
|
} catch (Exception e) { |
||||
|
log.info("id :{},消费者线程读取数据后解析失败", id); |
||||
|
e.printStackTrace(); |
||||
|
} |
||||
|
} |
||||
|
// 提交偏移量 |
||||
|
try { |
||||
|
consumer.commitSync(); |
||||
|
} catch (Exception e) { |
||||
|
log.error("提交偏移量失败", e); |
||||
|
} |
||||
|
|
||||
|
if (isBreak) { |
||||
|
consumer.close(); |
||||
|
Constant.CONSUMER_MAP.remove(objectKey); |
||||
|
Constant.TIME_OUT_MAP.remove(objectKey); |
||||
|
stringRedisTemplate.delete(objectKey); |
||||
|
} |
||||
|
} |
||||
|
consumer.close(); |
||||
|
log.info("id :{},消费者线程已经关闭", id); |
||||
|
Constant.CONSUMER_MAP.remove(objectKey); |
||||
|
Constant.TIME_OUT_MAP.remove(objectKey); |
||||
|
try { |
||||
|
stringRedisTemplate.delete(objectKey); |
||||
|
} catch (Exception e) { |
||||
|
throw new RuntimeException(e); |
||||
|
} |
||||
|
} catch (IllegalStateException illegalStateException) { |
||||
|
//这里的异常没有打印,考虑超时关闭退出的功能 |
||||
|
log.error("检测到消费者被关闭,任务:" + id + "的\t" + name + "\t消费线程退出"); |
||||
|
} catch (Throwable throwable) { |
||||
|
throwable.printStackTrace(); |
||||
|
log.error("kafka消费线程发生异常" + throwable.getMessage()); |
||||
|
Map result = new HashMap(32); |
||||
|
result.put("status", 2); |
||||
|
result.put("results", ""); |
||||
|
result.put("message", "未知异常"); |
||||
|
parse.put("result", result); |
||||
|
String message = JSON.toJSONString(parse); |
||||
|
try { |
||||
|
QueueUtil.sendQueue.put(message); |
||||
|
} catch (InterruptedException ex) { |
||||
|
ex.printStackTrace(); |
||||
|
} |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
public static void main(String[] args) { |
||||
|
|
||||
|
String json = ""; |
||||
|
Object parse = JSON.parse(json); |
||||
|
|
||||
|
} |
||||
|
} |
@ -0,0 +1,73 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.service; |
||||
|
|
||||
|
import com.alibaba.fastjson.JSON; |
||||
|
import com.bfd.crawl.kafkahandler.config.Constant; |
||||
|
import com.bfd.crawl.kafkahandler.util.QueueUtil; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.apache.kafka.clients.admin.AdminClient; |
||||
|
import org.apache.kafka.clients.admin.AdminClientConfig; |
||||
|
import org.apache.kafka.clients.admin.DescribeTopicsResult; |
||||
|
import org.apache.kafka.clients.admin.TopicDescription; |
||||
|
import org.apache.kafka.clients.consumer.ConsumerRecord; |
||||
|
import org.apache.kafka.clients.consumer.ConsumerRecords; |
||||
|
import org.apache.kafka.clients.consumer.KafkaConsumer; |
||||
|
import org.springframework.beans.factory.annotation.Autowired; |
||||
|
import org.springframework.data.redis.core.StringRedisTemplate; |
||||
|
import org.springframework.scheduling.annotation.Async; |
||||
|
import org.springframework.stereotype.Service; |
||||
|
|
||||
|
import javax.annotation.Resource; |
||||
|
import java.util.*; |
||||
|
import java.util.concurrent.ExecutionException; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:ConsumerHandlerService |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/8/22 16:54 |
||||
|
*/ |
||||
|
@Service |
||||
|
@Slf4j |
||||
|
public class ConsumerHandlerService { |
||||
|
@Autowired |
||||
|
private ConsumerHandler consumerHandler; |
||||
|
|
||||
|
@Async |
||||
|
public void conumerHandler(Integer id, String name, Map admin, Map parse, Map output, String objectKey) { |
||||
|
Constant.CONSUMER_RUNTIME_MAP.put(id, true); |
||||
|
String topic = (String) admin.get("topic"); |
||||
|
String bootstrapServers = (String) admin.get("bootstrapServers"); |
||||
|
int totalPartitions = getTotalPartitions(topic, bootstrapServers); |
||||
|
String groupId = UUID.randomUUID().toString(); |
||||
|
for (int i = 0; i < totalPartitions; i++) { |
||||
|
consumerHandler.creatConsumer(id, name, admin, parse, output, objectKey, groupId); |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
private static Integer getTotalPartitions(String topicName, String bootstrapServers) { |
||||
|
int totalPartitions = 1; |
||||
|
// Kafka管理员客户端的配置 |
||||
|
Properties adminClientConfig = new Properties(); |
||||
|
adminClientConfig.put(AdminClientConfig.BOOTSTRAP_SERVERS_CONFIG, bootstrapServers); |
||||
|
|
||||
|
// 创建Kafka管理员客户端 |
||||
|
try (AdminClient adminClient = AdminClient.create(adminClientConfig)) { |
||||
|
// 获取指定Topic的描述信息 |
||||
|
DescribeTopicsResult describeTopicsResult = adminClient.describeTopics(Collections.singletonList(topicName)); |
||||
|
Map<String, TopicDescription> topicDescriptionMap = describeTopicsResult.all().get(); |
||||
|
|
||||
|
// 获取并打印指定Topic的分区数 |
||||
|
TopicDescription topicDescription = topicDescriptionMap.get(topicName); |
||||
|
if (topicDescription != null) { |
||||
|
totalPartitions = topicDescription.partitions().size(); |
||||
|
log.info("Topic: " + topicName + ", Total Partitions: " + totalPartitions); |
||||
|
} else { |
||||
|
log.info("Topic: " + topicName + " not found."); |
||||
|
} |
||||
|
} catch (InterruptedException | ExecutionException e) { |
||||
|
e.printStackTrace(); |
||||
|
} |
||||
|
return totalPartitions; |
||||
|
} |
||||
|
} |
@ -0,0 +1,93 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.service; |
||||
|
|
||||
|
import com.alibaba.fastjson.JSON; |
||||
|
import com.alibaba.fastjson.JSONObject; |
||||
|
import com.bfd.crawl.kafkahandler.config.Constant; |
||||
|
import com.bfd.crawl.kafkahandler.util.DataUtil; |
||||
|
import com.bfd.crawl.kafkahandler.util.QueueUtil; |
||||
|
import com.google.gson.Gson; |
||||
|
import com.google.gson.GsonBuilder; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.apache.kafka.clients.producer.KafkaProducer; |
||||
|
import org.apache.kafka.clients.producer.ProducerRecord; |
||||
|
import org.springframework.scheduling.annotation.Async; |
||||
|
import org.springframework.stereotype.Service; |
||||
|
|
||||
|
import java.util.HashMap; |
||||
|
import java.util.List; |
||||
|
import java.util.Map; |
||||
|
import java.util.Properties; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:ProducterHandlerService |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/8/28 14:40 |
||||
|
*/ |
||||
|
@Service |
||||
|
@Slf4j |
||||
|
public class ProducterHandlerService { |
||||
|
|
||||
|
@Async("producterExecutor") |
||||
|
public void producterHandler(Integer id, String name, Map admin, Map dataMap, Map parse) { |
||||
|
Gson gson = new GsonBuilder().serializeNulls().create(); |
||||
|
log.info("任务ID:" + id + "," + name + "生产者线程启动"); |
||||
|
String bootstrapServers = (String) admin.get("bootstrapServers"); |
||||
|
String topic = (String) admin.get("topic"); |
||||
|
String message = ""; |
||||
|
Properties props = new Properties(); |
||||
|
props.put("bootstrap.servers", bootstrapServers); |
||||
|
props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); |
||||
|
props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); |
||||
|
props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer"); |
||||
|
props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer"); |
||||
|
props.put("request.timeout.ms", "60000"); |
||||
|
props.put("session.timeout.ms", "60000"); |
||||
|
props.put("max.request.size", "47185920"); |
||||
|
|
||||
|
KafkaProducer<String, String> producer = new KafkaProducer<String, String>(props); |
||||
|
|
||||
|
// 设置请求体,包含要写入的文档数据 |
||||
|
Map<String, Object> document = new HashMap<String, Object>(16); |
||||
|
//获取写入字段 |
||||
|
List<Map<String, Object>> form = (List<Map<String, Object>>) admin.get(Constant.FORM); |
||||
|
for (Map<String, Object> map : form) { |
||||
|
//字段名称 |
||||
|
String field = (String) map.get(Constant.FIELD); |
||||
|
//数据源path公式 |
||||
|
String valuePath = (String) map.get(Constant.VALUE); |
||||
|
//根据path去数据源data下获取字段值 |
||||
|
Object value = DataUtil.getValue(valuePath, dataMap); |
||||
|
//字段为空不处理 |
||||
|
// if (value == null) { |
||||
|
// continue; |
||||
|
// } |
||||
|
//判断是不是 String类型 ---暂时先判断String类型 |
||||
|
//其他类型直接进行赋值,--暂定,因为前端传过来的字段是根据我们应用进行定义的,类型基本一致 |
||||
|
document.put(field, value); |
||||
|
} |
||||
|
// message = JSONObject.toJSONString(document); |
||||
|
document.put("isLast", 1); |
||||
|
message = gson.toJson(document); |
||||
|
try { |
||||
|
ProducerRecord<String, String> producerRecord = new ProducerRecord<String, String>(topic, message); |
||||
|
producer.send(producerRecord); |
||||
|
producer.flush(); |
||||
|
producer.close(); |
||||
|
} catch (Throwable throwable) { |
||||
|
throwable.printStackTrace(); |
||||
|
log.error("kafka生产线程发生异常" + throwable.getMessage()); |
||||
|
} |
||||
|
Map result = new HashMap(32); |
||||
|
result.put("results", message); |
||||
|
result.put("status", 1); |
||||
|
parse.put("result", result); |
||||
|
String nextTask = JSON.toJSONString(parse); |
||||
|
try { |
||||
|
QueueUtil.sendQueue.put(nextTask); |
||||
|
} catch (InterruptedException e) { |
||||
|
e.printStackTrace(); |
||||
|
} |
||||
|
} |
||||
|
} |
@ -0,0 +1,102 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.service; |
||||
|
|
||||
|
import com.alibaba.fastjson.JSON; |
||||
|
import com.alibaba.fastjson.JSONObject; |
||||
|
import com.alibaba.fastjson.JSONPath; |
||||
|
import com.bfd.crawl.kafkahandler.config.Constant; |
||||
|
import com.bfd.crawl.kafkahandler.util.QueueUtil; |
||||
|
import com.bfd.crawl.kafkahandler.util.StringUtil; |
||||
|
import com.bfd.util.PauseTool; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.springframework.beans.factory.annotation.Autowired; |
||||
|
import org.springframework.boot.ApplicationArguments; |
||||
|
import org.springframework.boot.ApplicationRunner; |
||||
|
import org.springframework.data.redis.core.StringRedisTemplate; |
||||
|
import org.springframework.scheduling.annotation.Async; |
||||
|
import org.springframework.stereotype.Service; |
||||
|
|
||||
|
import javax.annotation.Resource; |
||||
|
import java.util.HashMap; |
||||
|
import java.util.Map; |
||||
|
import java.util.concurrent.TimeUnit; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:SchHandlerService |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/8/28 14:52 |
||||
|
*/ |
||||
|
|
||||
|
@Service |
||||
|
@Slf4j |
||||
|
public class SchHandlerService { |
||||
|
@Resource |
||||
|
private StringRedisTemplate stringRedisTemplate; |
||||
|
|
||||
|
@Autowired |
||||
|
private ConsumerHandlerService consumerHandlerService; |
||||
|
@Autowired |
||||
|
private ProducterHandlerService producterHandlerService; |
||||
|
@Autowired |
||||
|
private TimeOutHandlerService timeOutHandlerService; |
||||
|
|
||||
|
|
||||
|
@Async |
||||
|
public void run() { |
||||
|
//先执行对象的超时处理线程 |
||||
|
// timeOutHandlerService.handlerTimeOut(); |
||||
|
while (true) { |
||||
|
try { |
||||
|
String dataJson = QueueUtil.taskQueue.take(); |
||||
|
Map parse = (Map) JSON.parse(dataJson); |
||||
|
String objectKey = parse.get("scenes_id").toString(); |
||||
|
int id = (int) parse.get("scenes_id"); |
||||
|
String name = (String) parse.get("name"); |
||||
|
Map dataMap = (Map) parse.get("data"); |
||||
|
Map admin = (Map) parse.get("input"); |
||||
|
Map output = (Map) parse.get("output"); |
||||
|
int scenesId = (int) parse.get("scenes_id"); |
||||
|
int version = (int) parse.get("version"); |
||||
|
String pauseKey = scenesId + "_" + version; |
||||
|
if (!PauseTool.CACHE.containsKey(pauseKey)) { |
||||
|
log.info("流程:{}的版本:{}已失效,任务跳过", scenesId, version); |
||||
|
continue; |
||||
|
} |
||||
|
// dataMap.put("blueprint_id", parse.get("blueprint_id")); |
||||
|
// dataMap.put("scenes_id", parse.get("scenes_id")); |
||||
|
// dataMap.put("id", parse.get("id")); |
||||
|
// dataMap.put("transfer_id", parse.get("transfer_id")); |
||||
|
// dataMap.put("businessKey", parse.get("businessKey")); |
||||
|
int type = (int) admin.get("type"); |
||||
|
int timeOut = (int) admin.get("timeOut"); |
||||
|
|
||||
|
stringRedisTemplate.opsForHash().put(objectKey, "data", dataJson); |
||||
|
if (timeOut > 0) { |
||||
|
stringRedisTemplate.expire(objectKey, timeOut, TimeUnit.SECONDS); |
||||
|
} |
||||
|
//加入超时的相关信息,该信息针对生产与消费,要求前端进行判断,消费者默认是-1,生产者默认是600s |
||||
|
// if (!Constant.TIME_OUT_MAP.containsKey(objectKey)) { |
||||
|
// Map timeOutMap = new HashMap(32); |
||||
|
// timeOutMap.put("timeOut", timeOut); |
||||
|
// timeOutMap.put("startTime", System.currentTimeMillis()); |
||||
|
// Constant.TIME_OUT_MAP.put(objectKey, timeOutMap); |
||||
|
// }else { |
||||
|
// |
||||
|
// } |
||||
|
if (type == Constant.PRODUCER_TYPE) { |
||||
|
producterHandlerService.producterHandler(id, name, admin, dataMap, parse); |
||||
|
} else { |
||||
|
//根据传入的条件判断当前是否有一样的程序在执行,若有则不会重复启动对应的处理线程,生产者不需要这个判断,每次收到既生产就可以 |
||||
|
if (!Constant.CONSUMER_MAP.containsKey(objectKey)) { |
||||
|
consumerHandlerService.conumerHandler(id, name, admin, parse, output, objectKey); |
||||
|
} |
||||
|
} |
||||
|
} catch (Throwable e) { |
||||
|
e.printStackTrace(); |
||||
|
log.error("工作线程发生异常:{}", e.getMessage()); |
||||
|
} |
||||
|
|
||||
|
} |
||||
|
} |
||||
|
} |
@ -0,0 +1,61 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.service; |
||||
|
|
||||
|
import com.alibaba.fastjson.JSON; |
||||
|
import com.bfd.crawl.kafkahandler.util.QueueUtil; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.springframework.beans.factory.annotation.Autowired; |
||||
|
import org.springframework.beans.factory.annotation.Value; |
||||
|
import org.springframework.kafka.core.KafkaTemplate; |
||||
|
import org.springframework.scheduling.annotation.Async; |
||||
|
import org.springframework.stereotype.Service; |
||||
|
|
||||
|
import java.util.Map; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:SendService |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/7/31 17:53 |
||||
|
*/ |
||||
|
@Slf4j |
||||
|
@Service |
||||
|
public class SendService { |
||||
|
@Value("${send.topic}") |
||||
|
private String topic; |
||||
|
@Autowired |
||||
|
private KafkaTemplate kafkaTemplate; |
||||
|
|
||||
|
@Async() |
||||
|
void sendToKafka() { |
||||
|
while (true) { |
||||
|
if (QueueUtil.sendQueue.size() > 0) { |
||||
|
String message = null; |
||||
|
String version = null; |
||||
|
String id = null; |
||||
|
try { |
||||
|
message = QueueUtil.sendQueue.take(); |
||||
|
Map parse = (Map) JSON.parse(message); |
||||
|
version = parse.get("version").toString(); |
||||
|
id = parse.get("id").toString(); |
||||
|
} catch (InterruptedException e) { |
||||
|
throw new RuntimeException(e); |
||||
|
} |
||||
|
try { |
||||
|
kafkaTemplate.send(topic, message); |
||||
|
log.info("id:{},version:{}数据已发出", id, version); |
||||
|
} catch (Exception e) { |
||||
|
log.info("id:{},version:{}数据发送异常:{}", id, version, e.getMessage()); |
||||
|
e.printStackTrace(); |
||||
|
} |
||||
|
} else { |
||||
|
log.info("任务队列为空,休眠3秒"); |
||||
|
try { |
||||
|
Thread.sleep(3000); |
||||
|
} catch (InterruptedException e) { |
||||
|
e.printStackTrace(); |
||||
|
} |
||||
|
} |
||||
|
} |
||||
|
} |
||||
|
} |
@ -0,0 +1,60 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.service; |
||||
|
|
||||
|
import com.bfd.crawl.kafkahandler.util.QueueUtil; |
||||
|
import com.bfd.util.PauseTool; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.springframework.beans.factory.annotation.Autowired; |
||||
|
import org.springframework.beans.factory.annotation.Value; |
||||
|
import org.springframework.boot.ApplicationArguments; |
||||
|
import org.springframework.boot.ApplicationRunner; |
||||
|
import org.springframework.data.redis.core.StringRedisTemplate; |
||||
|
import org.springframework.stereotype.Service; |
||||
|
|
||||
|
import javax.annotation.Resource; |
||||
|
import java.util.Set; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:StartServcie |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/7/31 17:14 |
||||
|
*/ |
||||
|
@Service |
||||
|
@Slf4j |
||||
|
public class StartServcie implements ApplicationRunner { |
||||
|
@Autowired |
||||
|
private SchHandlerService schHandlerService; |
||||
|
@Value("${zookeeper.connection-string}") |
||||
|
private String connectionString; |
||||
|
@Value("${zookeeper.publish-node}") |
||||
|
private String nodePath; |
||||
|
@Resource |
||||
|
private StringRedisTemplate stringRedisTemplate; |
||||
|
@Value("${thread.send}") |
||||
|
private int sendNumber; |
||||
|
@Autowired |
||||
|
private SendService sendService; |
||||
|
|
||||
|
@Override |
||||
|
public void run(ApplicationArguments args) throws Exception { |
||||
|
PauseTool pauseTool = new PauseTool(); |
||||
|
pauseTool.initializeRedisCache(stringRedisTemplate); |
||||
|
pauseTool.setupZookeeperListener(connectionString, nodePath); |
||||
|
for (int i = 0; i < sendNumber; i++) { |
||||
|
sendService.sendToKafka(); |
||||
|
} |
||||
|
schHandlerService.run(); |
||||
|
try { |
||||
|
Set<String> keys = stringRedisTemplate.keys("*"); |
||||
|
for (String key : keys) { |
||||
|
String data = (String) stringRedisTemplate.opsForHash().get(key, "data"); |
||||
|
QueueUtil.taskQueue.put(data); |
||||
|
} |
||||
|
} catch (Exception e) { |
||||
|
|
||||
|
} |
||||
|
|
||||
|
|
||||
|
} |
||||
|
} |
@ -0,0 +1,66 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.service; |
||||
|
|
||||
|
import com.bfd.crawl.kafkahandler.config.Constant; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.springframework.data.redis.core.StringRedisTemplate; |
||||
|
import org.springframework.scheduling.annotation.Async; |
||||
|
import org.springframework.stereotype.Service; |
||||
|
|
||||
|
import javax.annotation.Resource; |
||||
|
import java.util.Map; |
||||
|
import java.util.concurrent.ConcurrentHashMap; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:TimeOutHandlerService |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/8/29 17:54 |
||||
|
*/ |
||||
|
@Service |
||||
|
@Slf4j |
||||
|
public class TimeOutHandlerService { |
||||
|
@Resource |
||||
|
private StringRedisTemplate stringRedisTemplate; |
||||
|
|
||||
|
@Async("asyncExecutor") |
||||
|
public void handlerTimeOut() { |
||||
|
while (true) { |
||||
|
try { |
||||
|
if (Constant.TIME_OUT_MAP.size() > 0) { |
||||
|
ConcurrentHashMap.KeySetView<String, Map> strings = Constant.TIME_OUT_MAP.keySet(); |
||||
|
for (String string : strings) { |
||||
|
Map map = Constant.TIME_OUT_MAP.get(string); |
||||
|
int timeOut = (int) map.get("timeOut"); |
||||
|
long startTime = (long) map.get("startTime"); |
||||
|
long timeConsuming = (System.currentTimeMillis() - startTime) / 1000; |
||||
|
//如果超时时间为-1或者当前时间减开始时间小于超时时间,不处理 |
||||
|
if (timeOut == -1 || timeConsuming < timeOut) { |
||||
|
continue; |
||||
|
} |
||||
|
if (Constant.CONSUMER_MAP.containsKey(string)) { |
||||
|
Constant.CONSUMER_MAP.get(string).close(); |
||||
|
|
||||
|
} |
||||
|
if (Constant.PRODUCER_MAP.containsKey(string)) { |
||||
|
Constant.PRODUCER_MAP.get(string).close(); |
||||
|
} |
||||
|
try { |
||||
|
stringRedisTemplate.delete(string); |
||||
|
} catch (Exception e) { |
||||
|
|
||||
|
} |
||||
|
|
||||
|
} |
||||
|
Thread.sleep(30000); |
||||
|
} else { |
||||
|
log.info("当前队列为空休眠10s"); |
||||
|
Thread.sleep(10000); |
||||
|
} |
||||
|
} catch (Throwable throwable) { |
||||
|
throwable.printStackTrace(); |
||||
|
log.error("计时线程发生异常" + throwable.getMessage()); |
||||
|
} |
||||
|
} |
||||
|
} |
||||
|
} |
@ -0,0 +1,66 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.service; |
||||
|
|
||||
|
import com.alibaba.fastjson.JSON; |
||||
|
import com.alibaba.fastjson.JSONObject; |
||||
|
import com.bfd.crawl.kafkahandler.config.Constant; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
import org.apache.curator.framework.CuratorFramework; |
||||
|
import org.apache.curator.framework.recipes.cache.NodeCache; |
||||
|
import org.apache.curator.framework.recipes.cache.NodeCacheListener; |
||||
|
import org.springframework.beans.factory.annotation.Autowired; |
||||
|
import org.springframework.beans.factory.annotation.Value; |
||||
|
import org.springframework.data.redis.core.StringRedisTemplate; |
||||
|
import org.springframework.stereotype.Component; |
||||
|
|
||||
|
import javax.annotation.PostConstruct; |
||||
|
import javax.annotation.Resource; |
||||
|
|
||||
|
/** |
||||
|
* @author jian.mao |
||||
|
* @date 2024年4月17日 |
||||
|
* @description |
||||
|
*/ |
||||
|
@Component |
||||
|
@Slf4j |
||||
|
public class ZookeeperNodeMonitor { |
||||
|
|
||||
|
@Resource |
||||
|
private StringRedisTemplate stringRedisTemplate; |
||||
|
@Autowired |
||||
|
private CuratorFramework curatorFramework; |
||||
|
@Value("${zookeeper.publish-node}") |
||||
|
private String nodePath; |
||||
|
|
||||
|
@PostConstruct |
||||
|
public void init() { |
||||
|
try { |
||||
|
// 创建节点监听器 |
||||
|
NodeCache nodeCache = new NodeCache(curatorFramework, nodePath); |
||||
|
nodeCache.start(); |
||||
|
|
||||
|
// 监听节点变化 |
||||
|
nodeCache.getListenable().addListener(new NodeCacheListener() { |
||||
|
@Override |
||||
|
public void nodeChanged() throws Exception { |
||||
|
byte[] data = nodeCache.getCurrentData().getData(); |
||||
|
String nodeData = new String(data); |
||||
|
log.info("Node data changed: " + nodeData); |
||||
|
JSONObject jsonObject = JSON.parseObject(nodeData); |
||||
|
int scenesId = jsonObject.getIntValue("scenes_id"); |
||||
|
Constant.CONSUMER_RUNTIME_MAP.put(scenesId, false); |
||||
|
if (Constant.PRODUCER_MAP.containsKey(nodeData)) { |
||||
|
Constant.PRODUCER_MAP.get(nodeData).close(); |
||||
|
Constant.PRODUCER_MAP.remove(scenesId); |
||||
|
} |
||||
|
try { |
||||
|
stringRedisTemplate.delete(nodeData); |
||||
|
} catch (Exception e) { |
||||
|
|
||||
|
} |
||||
|
} |
||||
|
}); |
||||
|
} catch (Exception e) { |
||||
|
e.printStackTrace(); |
||||
|
} |
||||
|
} |
||||
|
} |
@ -0,0 +1,65 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.util; |
||||
|
|
||||
|
import java.util.Map; |
||||
|
|
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
|
||||
|
import com.alibaba.fastjson.JSON; |
||||
|
import com.alibaba.fastjson.JSONObject; |
||||
|
import com.alibaba.fastjson.JSONPath; |
||||
|
import com.bfd.crawl.kafkahandler.config.Constant; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:DataUtil |
||||
|
* @version:1.0 |
||||
|
* @description: 获取dataValue的值 |
||||
|
* @Date:2023/11/1 9:54 |
||||
|
*/ |
||||
|
@Slf4j |
||||
|
public class DataUtil { |
||||
|
/** |
||||
|
* @param key 传入的key |
||||
|
* @param dataMap 数据map |
||||
|
* @return 根据传入的参数进行判断解析,返回正确的dataValue |
||||
|
*/ |
||||
|
public static Object getValue(String key, Map dataMap) { |
||||
|
try { |
||||
|
//公式为空直接就返回 |
||||
|
if (key.equals(Constant.EMPTY)) { |
||||
|
return Constant.EMPTY; |
||||
|
} |
||||
|
if (!key.contains(Constant.NOT_KEY)) { |
||||
|
return key; |
||||
|
} |
||||
|
Object dataValue; |
||||
|
String isJson = "#json#"; |
||||
|
if (key.contains(isJson)) { |
||||
|
//进行第一次拆分,获取#json#前面的部分 |
||||
|
String[] keySplit = key.split(isJson); |
||||
|
String firstDataKey = keySplit[0]; |
||||
|
String[] firstDataKeySplit = firstDataKey.split(":"); |
||||
|
//取出前半部分对应的JSON数据并转换为JSONObject |
||||
|
String dataJson = (String) dataMap.get(firstDataKeySplit[0]); |
||||
|
JSONObject dataJsonObject = JSON.parseObject(dataJson); |
||||
|
//根据key的后半部分取出对应JSONObject中的值 |
||||
|
String firstDataKeyJson = (String) JSONPath.eval(dataJsonObject, firstDataKeySplit[1]); |
||||
|
String secDataKey = keySplit[1]; |
||||
|
JSONObject firstDataJsonObject = JSON.parseObject(firstDataKeyJson); |
||||
|
dataValue = JSONPath.eval(firstDataJsonObject, secDataKey); |
||||
|
return dataValue; |
||||
|
} |
||||
|
String[] keySplit = key.split(":"); |
||||
|
String jsonPath = keySplit[1]; |
||||
|
String dataJson = (String) dataMap.get(keySplit[0]); |
||||
|
JSONObject dataJsonObject = JSON.parseObject(dataJson); |
||||
|
dataValue = JSONPath.eval(dataJsonObject, jsonPath); |
||||
|
return dataValue; |
||||
|
} catch (Exception e) { |
||||
|
// TODO: handle exception |
||||
|
log.error("jsonpath公式取值异常,", e); |
||||
|
return null; |
||||
|
} |
||||
|
|
||||
|
} |
||||
|
} |
@ -0,0 +1,16 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.util; |
||||
|
|
||||
|
import java.util.concurrent.LinkedBlockingDeque; |
||||
|
|
||||
|
/** |
||||
|
* @author:jinming |
||||
|
* @className:QueueUtil |
||||
|
* @version:1.0 |
||||
|
* @description: |
||||
|
* @Date:2023/7/13 15:00 |
||||
|
*/ |
||||
|
public class QueueUtil { |
||||
|
public static LinkedBlockingDeque<String> taskQueue = new LinkedBlockingDeque<String>(); |
||||
|
|
||||
|
public static LinkedBlockingDeque<String> sendQueue = new LinkedBlockingDeque<String>(); |
||||
|
} |
@ -0,0 +1,147 @@ |
|||||
|
package com.bfd.crawl.kafkahandler.util; |
||||
|
|
||||
|
|
||||
|
import com.alibaba.fastjson.JSONObject; |
||||
|
import lombok.extern.slf4j.Slf4j; |
||||
|
|
||||
|
import java.security.MessageDigest; |
||||
|
import java.util.HashSet; |
||||
|
import java.util.Set; |
||||
|
import java.util.regex.Matcher; |
||||
|
import java.util.regex.Pattern; |
||||
|
|
||||
|
/** |
||||
|
* @author jinming |
||||
|
* @version 1.0 |
||||
|
* @className StringUtile |
||||
|
* @Date 2022/1/21 11:46 |
||||
|
*/ |
||||
|
@Slf4j |
||||
|
public class StringUtil { |
||||
|
public static boolean hasValue(String str) { |
||||
|
return str != null && !"".equals(str.trim()); |
||||
|
} |
||||
|
|
||||
|
public static String getRegexGroup(String regex, String str, int id) { |
||||
|
String resultStr = ""; |
||||
|
if (hasValue(str)) { |
||||
|
Pattern p = Pattern.compile(regex); |
||||
|
Matcher m = p.matcher(str); |
||||
|
if (m.find()) { |
||||
|
resultStr = m.group(id); |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
if ("".equals(resultStr)) { |
||||
|
} |
||||
|
|
||||
|
return resultStr; |
||||
|
} |
||||
|
|
||||
|
/** |
||||
|
* |
||||
|
* @param str1 原始字符串1 |
||||
|
* @param str2 原始字符串2 |
||||
|
* @return 并集后的新字符串 |
||||
|
*/ |
||||
|
public static String stringUnion(String str1, String str2) { |
||||
|
// 将两个字符串的字符合并到一个集合中,自动去除重复字符 |
||||
|
Set<Character> unionSet = new HashSet<>(); |
||||
|
|
||||
|
for (char c : str1.toCharArray()) { |
||||
|
unionSet.add(c); |
||||
|
} |
||||
|
|
||||
|
for (char c : str2.toCharArray()) { |
||||
|
unionSet.add(c); |
||||
|
} |
||||
|
|
||||
|
// 将集合转换为字符串 |
||||
|
StringBuilder result = new StringBuilder(); |
||||
|
for (char c : unionSet) { |
||||
|
result.append(c); |
||||
|
} |
||||
|
|
||||
|
return result.toString(); |
||||
|
} |
||||
|
|
||||
|
|
||||
|
|
||||
|
// 递归函数来插入值 |
||||
|
public static void insertValue(JSONObject jsonObject, String[] path, String value) { |
||||
|
if (path.length == 0) { |
||||
|
return; |
||||
|
} |
||||
|
String key = path[0]; |
||||
|
|
||||
|
if (path.length == 1) { |
||||
|
jsonObject.put(key, value); |
||||
|
} else { |
||||
|
JSONObject subObject = jsonObject.getJSONObject(key); |
||||
|
if (subObject == null) { |
||||
|
subObject = new JSONObject(); |
||||
|
jsonObject.put(key, subObject); |
||||
|
} |
||||
|
|
||||
|
String[] newPath = new String[path.length - 1]; |
||||
|
System.arraycopy(path, 1, newPath, 0, newPath.length); |
||||
|
insertValue(subObject, newPath, value); |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
public static Set<String> getEmailAddress(String message) { |
||||
|
Set<String> emailList = new HashSet<>(); |
||||
|
Pattern pattern = Pattern.compile("\\w+\\.?\\w+\\@\\w+\\.\\w+"); |
||||
|
Matcher m = pattern.matcher(message); |
||||
|
while (m.find()) { |
||||
|
emailList.add(m.group(0)); |
||||
|
} |
||||
|
return emailList; |
||||
|
} |
||||
|
|
||||
|
public static String getMd5(String string) { |
||||
|
try { |
||||
|
MessageDigest md5 = MessageDigest.getInstance("MD5"); |
||||
|
byte[] bs = md5.digest(string.getBytes("UTF-8")); |
||||
|
StringBuilder sb = new StringBuilder(40); |
||||
|
for (byte x : bs) { |
||||
|
if ((x & 0xff) >> 4 == 0) { |
||||
|
sb.append("0").append(Integer.toHexString(x & 0xff)); |
||||
|
} else { |
||||
|
sb.append(Integer.toHexString(x & 0xff)); |
||||
|
} |
||||
|
} |
||||
|
return sb.toString(); |
||||
|
} catch (Exception e) { |
||||
|
//LOG.error("获取md5异常", e); |
||||
|
return "nceaform" + System.currentTimeMillis(); |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
public static String removeAllHtmlTags(String str) { |
||||
|
return hasValue(str) ? str.replaceAll("<[^<>]+?>", "") : ""; |
||||
|
} |
||||
|
|
||||
|
public static String getRegexGroup(Pattern regex, String str, int id) { |
||||
|
String resultStr = ""; |
||||
|
if (hasValue(str)) { |
||||
|
Matcher m = regex.matcher(str); |
||||
|
if (m.find()) { |
||||
|
resultStr = m.group(id); |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
if ("".equals(resultStr)) { |
||||
|
log.error(regex + " parser error!"); |
||||
|
} |
||||
|
|
||||
|
return resultStr; |
||||
|
} |
||||
|
|
||||
|
public static String getStrByPattern(String str, String regex) { |
||||
|
Pattern pattern = Pattern.compile(regex); |
||||
|
Matcher m = pattern.matcher(str); |
||||
|
return m.find() ? m.group(0) : ""; |
||||
|
} |
||||
|
|
||||
|
} |
@ -0,0 +1,57 @@ |
|||||
|
server: |
||||
|
port: 13188 |
||||
|
spring: |
||||
|
application: |
||||
|
name: 过滤器 |
||||
|
boot: |
||||
|
admin: |
||||
|
client: |
||||
|
health: |
||||
|
timeout: 10s |
||||
|
url: http://172.16.12.55:8001 |
||||
|
instance: |
||||
|
service-base-url: http://10.10.144.49:9080 |
||||
|
kafka: |
||||
|
bootstrap-servers: 172.18.1.146:9092,172.18.1.147:9092,172.18.1.148:9092 |
||||
|
producer: |
||||
|
retries: 3 |
||||
|
acks: all |
||||
|
batch-size: 4096 |
||||
|
buffer-memory: 102476800 |
||||
|
key-serializer: org.apache.kafka.common.serialization.StringSerializer |
||||
|
value-serializer: org.apache.kafka.common.serialization.StringSerializer |
||||
|
properties: |
||||
|
max.request.size: 47185920 |
||||
|
|
||||
|
redis: |
||||
|
host: 172.18.1.157 |
||||
|
port: 6379 |
||||
|
timeout: 10000 |
||||
|
database: 6 |
||||
|
jedis: |
||||
|
pool: |
||||
|
max-active: 8 # ???????????????????? |
||||
|
max-wait: 800 # ??????????????????????? |
||||
|
max-idle: 8 # ??????????? |
||||
|
min-idle: 2 # ??????????? |
||||
|
logging: |
||||
|
file: |
||||
|
path: ./logs |
||||
|
thread: |
||||
|
handler: 1 |
||||
|
send: 1 |
||||
|
|
||||
|
zookeeper: |
||||
|
connection-string: 172.18.1.146:2181,172.18.1.147:2181,172.18.1.148:2181 |
||||
|
publish-node: /analyze |
||||
|
management: |
||||
|
endpoints: |
||||
|
web: |
||||
|
exposure: |
||||
|
include: "*" |
||||
|
endpoint: |
||||
|
health: |
||||
|
show-details: always |
||||
|
|
||||
|
send: |
||||
|
topic: analyze12321 |
@ -0,0 +1,36 @@ |
|||||
|
<configuration> |
||||
|
<!-- 属性文件:在properties文件中找到对应的配置项 --> |
||||
|
<springProperty scope="context" name="logging.file.path" source="logging.file.path"/> |
||||
|
<springProperty scope="context" name="logging.level" source="logging.level"/> |
||||
|
<!-- 默认的控制台日志输出,一般生产环境都是后台启动,这个没太大作用 --> |
||||
|
<appender name="STDOUT" |
||||
|
class="ch.qos.logback.core.ConsoleAppender"> |
||||
|
<encoder class="ch.qos.logback.classic.encoder.PatternLayoutEncoder"> |
||||
|
<pattern>%d{yyyy-MM-dd HH:mm:ss.SSS} [%thread] %line %-5level %logger{50} - %msg%n</pattern> |
||||
|
</encoder> |
||||
|
</appender> |
||||
|
|
||||
|
<appender name="GLMAPPER-LOGGERONE" |
||||
|
class="ch.qos.logback.core.rolling.RollingFileAppender"> |
||||
|
<append>true</append> |
||||
|
<filter class="ch.qos.logback.classic.filter.ThresholdFilter"> |
||||
|
<level>${logging.level}</level> |
||||
|
</filter> |
||||
|
<file> |
||||
|
${logging.file.path}/kafkaHandler.log |
||||
|
</file> |
||||
|
<rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> |
||||
|
<FileNamePattern>${logging.file.path}/kafkaHandler.log.%d{yyyy-MM-dd}</FileNamePattern> |
||||
|
<MaxHistory>3</MaxHistory> |
||||
|
</rollingPolicy> |
||||
|
<encoder class="ch.qos.logback.classic.encoder.PatternLayoutEncoder"> |
||||
|
<pattern>%d{yyyy-MM-dd HH:mm:ss.SSS} [%thread] %line %-5level %logger{50} - %msg%n</pattern> |
||||
|
<charset>UTF-8</charset> |
||||
|
</encoder> |
||||
|
</appender> |
||||
|
|
||||
|
<root level="info"> |
||||
|
<appender-ref ref="GLMAPPER-LOGGERONE"/> |
||||
|
<appender-ref ref="STDOUT"/> |
||||
|
</root> |
||||
|
</configuration> |
@ -0,0 +1,13 @@ |
|||||
|
package com.bfd.crawl.kafkahandler; |
||||
|
|
||||
|
import org.junit.jupiter.api.Test; |
||||
|
import org.springframework.boot.test.context.SpringBootTest; |
||||
|
|
||||
|
@SpringBootTest |
||||
|
class KafkaHandlerApplicationTests { |
||||
|
|
||||
|
@Test |
||||
|
void contextLoads() { |
||||
|
} |
||||
|
|
||||
|
} |
Write
Preview
Loading…
Cancel
Save
Reference in new issue