From 37b4adda41f419b9d4e14c641385c0e6f0a59a83 Mon Sep 17 00:00:00 2001 From: markilue <745518019@qq.com> Date: Sun, 9 Oct 2022 13:36:50 +0800 Subject: [PATCH] =?UTF-8?q?=E9=A6=96=E6=AC=A1=E6=9B=B4=E6=96=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- Big_data_example/flume/pom.xml | 40 +- .../com/atguigu/flume/test/MQTTSource.java | 258 + .../com/atguigu/flume/test/MQTTSource1.java | 310 ++ .../com/atguigu/flume/test/MQTTSource2.java | 232 + Big_data_example/log/agent.log | 4354 +++++++++++++++++ .../rtgmall/controller/FirstController.java | 20 + .../rtgmall/controller/LoggerController.java | 55 + .../src/main/resources/logback.xml | 32 + .../rt-gmall-parent/gmall-realtime/pom.xml | 118 + .../gmall/realtime/app/dwd/BaseLogApp.java | 110 + .../gmall/realtime/utils/MyKafkaUtils.java | 29 + .../src/main/resources/log4j.properties | 5 + .../edu/error/log-error-2022-04-21.0.log | 1 - .../edu/info/log-info-2022-04-21.0.log | 47 - .../edu/info/log-info-2022-04-22.0.log | 200 - .../edu/info/log-info-2022-04-24.0.log | 444 -- .../edu/info/log-info-2022-04-25.0.log | 200 - .../edu/info/log-info-2022-04-27.0.log | 72 - .../service/service_edu/edu/log_error.log | 116 +- .../service/service_edu/edu/log_info.log | 202 +- .../backend/phm_backend/collecting/pom.xml | 19 + 21 files changed, 5621 insertions(+), 1243 deletions(-) create mode 100644 Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource.java create mode 100644 Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource1.java create mode 100644 Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource2.java create mode 100644 Big_data_example/rt-gmall-parent/gmall-logger/src/main/java/com/atguigu/rtgmall/controller/FirstController.java create mode 100644 Big_data_example/rt-gmall-parent/gmall-logger/src/main/java/com/atguigu/rtgmall/controller/LoggerController.java create mode 100644 Big_data_example/rt-gmall-parent/gmall-logger/src/main/resources/logback.xml create mode 100644 Big_data_example/rt-gmall-parent/gmall-realtime/pom.xml create mode 100644 Big_data_example/rt-gmall-parent/gmall-realtime/src/main/java/com/atguigu/gmall/realtime/app/dwd/BaseLogApp.java create mode 100644 Big_data_example/rt-gmall-parent/gmall-realtime/src/main/java/com/atguigu/gmall/realtime/utils/MyKafkaUtils.java create mode 100644 Big_data_example/rt-gmall-parent/gmall-realtime/src/main/resources/log4j.properties delete mode 100644 guli_parent/service/service_edu/edu/error/log-error-2022-04-21.0.log delete mode 100644 guli_parent/service/service_edu/edu/info/log-info-2022-04-21.0.log delete mode 100644 guli_parent/service/service_edu/edu/info/log-info-2022-04-22.0.log delete mode 100644 guli_parent/service/service_edu/edu/info/log-info-2022-04-24.0.log delete mode 100644 guli_parent/service/service_edu/edu/info/log-info-2022-04-25.0.log delete mode 100644 guli_parent/service/service_edu/edu/info/log-info-2022-04-27.0.log create mode 100644 phm_rotate/backend/phm_backend/collecting/pom.xml diff --git a/Big_data_example/flume/pom.xml b/Big_data_example/flume/pom.xml index c294c22..5318fb2 100644 --- a/Big_data_example/flume/pom.xml +++ b/Big_data_example/flume/pom.xml @@ -9,13 +9,40 @@ 1.0-SNAPSHOT + + + net.alchim31.maven + scala-maven-plugin + 3.4.6 + + + + + compile + testCompile + + + + + org.apache.maven.plugins - maven-compiler-plugin + maven-assembly-plugin + 3.0.0 - 7 - 7 + + jar-with-dependencies + + + + make-assembly + package + + single + + + @@ -27,6 +54,13 @@ 1.9.0 + + + org.eclipse.paho + org.eclipse.paho.client.mqttv3 + 1.2.2 + + diff --git a/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource.java b/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource.java new file mode 100644 index 0000000..c36b0af --- /dev/null +++ b/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource.java @@ -0,0 +1,258 @@ +package com.atguigu.flume.test; + +import org.apache.flume.*; +import org.apache.flume.conf.Configurable; +import org.apache.flume.event.EventBuilder; +import org.apache.flume.event.SimpleEvent; +import org.apache.flume.source.AbstractSource; +import org.eclipse.paho.client.mqttv3.*; + +import java.nio.charset.StandardCharsets; +import java.util.HashMap; +import java.util.Map; +import java.util.UUID; +import java.util.concurrent.TimeUnit; + + +/** + * @Author:huhy + * @DATE:Created on 2018/1/25 14:33 + * @Modified By: + * @Class Description: flume收集MQTT(Mosquitto)的数据。 + * 方法就是flume自定义source,source中来订阅(subscribe)MQTT + */ +public class MQTTSource extends AbstractSource implements EventDrivenSource, + Configurable { + + + /** + *源代码的初始化方法。上下文(context)包含了所有的内容。 + * Flume配置信息,可用于检索任何配置。 + *设置源所需的值。 + */ + @Override + public void configure(Context context) { + // TODO Auto-generated method stub + + } + + SimpleMqttClient client = null; + + /** + * Start any dependent systems and begin processing events. + * 启动任何依赖系统并开始处理事件。 + */ + @Override + public void start() { + // TODO Auto-generated method stub ||自动生成的方法 + // super.start(); + client = new SimpleMqttClient(); + client.runClient(); + } + + /** + * Stop processing events and shut any dependent systems down. + */ + @Override + public void stop() { + // TODO Auto-generated method stub || 自动生成的方法 + // super.stop(); + if (client != null) { + System.out.println("生命周期结束,关闭mqtt"); + client.closeConn(); + } + } + + + + public class SimpleMqttClient implements MqttCallback { + + MqttClient myClient; + MqttConnectOptions connOpt; + /** + * tcp连接 + */ +// String BROKER_URL = "tcp://192.168.118.202:1883"; + String BROKER_URL = "tcp://119.91.214.52:1883"; + + /** + * M2MIO_DOMAIN:ip + * M2MIO_STUFF:域 + */ + /*String M2MIO_DOMAIN = "192.168.1.19"; + String M2MIO_STUFF = "mytest";*/ + // topic + String M2MIO_THING = "37b0ce9f8e3643b48f23e5896d4d8c14"; + String M2MIO_USERNAME = "mqtt"; + String M2MIO_PASSWORD_MD5 ="mqtt"; + + //发布 订阅 + Boolean subscriber = true; + Boolean publisher = false; + + /** + * + * connectionLost This callback is invoked upon losing the MQTT + * connection. + * MQTT时调用connectionLost这个回调。 + * + */ + @Override + public void connectionLost(Throwable t) { + System.out.println("Connection lost!"); + // code to reconnect to the broker would go here if desired + } + + public void closeConn() { + if (myClient != null) { + if (myClient.isConnected()) { + try { + System.out.println("关闭mqtt"); + myClient.disconnect(); + } catch (MqttException e) { + // TODO Auto-generated catch block + e.printStackTrace(); + } + } + } + } + + /** + * + * deliveryComplete This callback is invoked when a message published by + * this client is successfully received by the broker. + * + */ + @Override + public void deliveryComplete(IMqttDeliveryToken token) { + // System.out.println("Pub complete" + new + // String(token.getMessage().getPayload())); + System.out.println("收到了一条数据"); + } + + /** + * 在接收到消息时调用messageArrived这个回调。 + * 一个订阅的主题。 + */ + @Override + public void messageArrived(String topic, MqttMessage message) + throws Exception { + System.out.println("getmessage..."); + + Map headers = new HashMap(); + headers.put("id", "123");//数据包带有唯一标示 + headers.put("time", String.valueOf(System.currentTimeMillis()));//设备的时间戳 + Event flumeEvent = EventBuilder.withBody(message.getPayload(), + headers); + try { + getChannelProcessor().processEvent(flumeEvent); + System.out.println("消息到达-------->"); + } catch (Exception e) { + // TODO: handle exception + e.printStackTrace(); + } + + } + + /** + * + * runClient The main functionality of this simple example. Create a + * MQTT client, connect to broker, pub/sub, disconnect. + * + */ + public void runClient() { + // setup MQTT Client + String clientID = M2MIO_THING; + connOpt = new MqttConnectOptions(); + + connOpt.setCleanSession(true); + connOpt.setKeepAliveInterval(3000); + connOpt.setUserName(M2MIO_USERNAME); + connOpt.setPassword(M2MIO_PASSWORD_MD5.toCharArray()); + + // Connect to Broker + try { + myClient = new MqttClient(BROKER_URL, clientID); + myClient.setCallback(this); + myClient.connect(connOpt); + System.out.println("连接成功"); + } catch (MqttException e) { + e.printStackTrace(); + System.exit(-1); + } + + System.out.println("Connected to " + BROKER_URL); + + // setup topic + // topics on m2m.io are in the form // + /** + * 方式一 + */ + /*String myTopic = M2MIO_DOMAIN + "/" + M2MIO_STUFF + "/" + + M2MIO_THING; + System.out.println("myTopic:" + myTopic); + MqttTopic topic = myClient.getTopic(myTopic);*/ + /** + * 方式二 + */ + String topic1= "/thing/"+M2MIO_THING+"/test_unchange_scada/data/post"; + MqttTopic topic = myClient.getTopic(topic1); + + // subscribe to topic if subscriber + if (subscriber) { + try { + System.out.println("开始订阅信息"); + int subQoS = 2; + // myClient.subscribe(myTopic, subQoS); + myClient.subscribe(topic1, subQoS); + + } catch (Exception e) { + e.printStackTrace(); + } + } + + // publish messages if publisher + if (publisher) { + System.out.println("开始发布消息"); + for (int i = 1; i <= 10; i++) { + String pubMsg = "{\"pubmsg\":" + i + "}"; + int pubQoS = 2; + MqttMessage message = new MqttMessage(pubMsg.getBytes()); + message.setQos(pubQoS); + message.setRetained(false); + + // Publish the message + System.out.println("Publishing to topic \"" + topic + + "\" qos " + pubQoS); + MqttDeliveryToken token = null; + try { + // publish message to broker + token = topic.publish(message); + // Wait until the message has been delivered to the + // broker + token.waitForCompletion(); + Thread.sleep(2000); + } catch (Exception e) { + e.printStackTrace(); + } + } + } + + // disconnect + try { + // wait to ensure subscribed messages are delivered + if (subscriber) { + while (true) { + Thread.sleep(5000); + } + } + // myClient.disconnect(); + } catch (Exception e) { + e.printStackTrace(); + } finally { + } + } + + } + +} diff --git a/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource1.java b/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource1.java new file mode 100644 index 0000000..910c1fe --- /dev/null +++ b/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource1.java @@ -0,0 +1,310 @@ +package com.atguigu.flume.test; + +import org.apache.flume.*; +import org.apache.flume.conf.Configurable; +import org.apache.flume.event.EventBuilder; +import org.apache.flume.event.SimpleEvent; +import org.apache.flume.source.AbstractSource; +import org.eclipse.paho.client.mqttv3.*; + +import java.nio.charset.StandardCharsets; +import java.util.HashMap; +import java.util.Map; +import java.util.UUID; +import java.util.concurrent.TimeUnit; + + +/** + * @Author:huhy + * @DATE:Created on 2018/1/25 14:33 + * @Modified By: + * @Class Description: flume收集MQTT(Mosquitto)的数据。 + * 方法就是flume自定义source,source中来订阅(subscribe)MQTT + */ +public class MQTTSource1 extends AbstractSource implements EventDrivenSource, + Configurable, PollableSource { + private String prefix; + + /** + *源代码的初始化方法。上下文(context)包含了所有的内容。 + * Flume配置信息,可用于检索任何配置。 + *设置源所需的值。 + */ + @Override + public void configure(Context context) { + // TODO Auto-generated method stub + prefix = context.getString("prefix","log--"); + + } + + SimpleMqttClient client = null; + + /** + * Start any dependent systems and begin processing events. + * 启动任何依赖系统并开始处理事件。 + */ + @Override + public void start() { + // TODO Auto-generated method stub ||自动生成的方法 + // super.start(); + client = new SimpleMqttClient(); + client.runClient(); + } + + /** + * Stop processing events and shut any dependent systems down. + */ + @Override + public void stop() { + // TODO Auto-generated method stub || 自动生成的方法 + // super.stop(); + if (client != null) { + System.out.println("生命周期结束"); + client.closeConn(); + } + } + + @Override + public Status process() throws EventDeliveryException { + + //当前生成数据太快了,休眠1秒钟 +// 可以用Thread.sleep(1000);但是要计算很麻烦,TimeUnit中有分别的小时,分秒的方法 + try { + TimeUnit.SECONDS.sleep(1); + } catch (InterruptedException e) { + e.printStackTrace(); + } + + Status status=null; + try { + // This try clause includes whatever Channel/Event operations you want to do + + // Receive new data + //采集数据,封装成event对象 + //创建getSomeData方法去获取数据 + Event e = getSomeData(); + + // Store the Event into this Source's associated Channel(s) + //将event对象交给ChannelProcessor进行处理 + getChannelProcessor().processEvent(e); + //正常处理,返回Status.READY + status = Status.READY; + } catch (Throwable t) { + // Log exception, handle individual exceptions as needed + //处理失败,返回Status.BACKOFF + status = Status.BACKOFF; +// // re-throw all Errors +// if (t instanceof Error) { +// throw (Error)t; +// } + } + return status; + } + + private Event getSomeData() { + //UUID.randomUUID()可以随机生成一个字符串,有固定的长度和格式 + String data= UUID.randomUUID().toString(); + String resultData = prefix+data; + //封装一个event类,使用SimpleEvent,在创建event时,内部已经创建好了header + SimpleEvent event = new SimpleEvent(); + //将字符串转化为字节数组放在event中 + event.setBody(resultData.getBytes(StandardCharsets.UTF_8)); + //可以创建source + event.getHeaders().put("author","Ding"); + return event; + } + + @Override + public long getBackOffSleepIncrement() { + return 1; + } + + @Override + public long getMaxBackOffSleepInterval() { + return 10; + } + + + public class SimpleMqttClient implements MqttCallback { + + MqttClient myClient; + MqttConnectOptions connOpt; + /** + * tcp连接 + */ + String BROKER_URL = "tcp://192.168.118.202:61613"; + /** + * M2MIO_DOMAIN:ip + * M2MIO_STUFF:域 + */ + /*String M2MIO_DOMAIN = "192.168.1.19"; + String M2MIO_STUFF = "mytest";*/ + // topic + String M2MIO_THING = "huhy"; + String M2MIO_USERNAME = "admin"; + String M2MIO_PASSWORD_MD5 ="password"; + + //发布 订阅 + Boolean subscriber = true; + Boolean publisher = false; + + /** + * + * connectionLost This callback is invoked upon losing the MQTT + * connection. + * MQTT时调用connectionLost这个回调。 + * + */ + @Override + public void connectionLost(Throwable t) { + System.out.println("Connection lost!"); + // code to reconnect to the broker would go here if desired + } + + public void closeConn() { + if (myClient != null) { + if (myClient.isConnected()) { + try { + myClient.disconnect(); + } catch (MqttException e) { + // TODO Auto-generated catch block + e.printStackTrace(); + } + } + } + } + + /** + * + * deliveryComplete This callback is invoked when a message published by + * this client is successfully received by the broker. + * + */ + @Override + public void deliveryComplete(IMqttDeliveryToken token) { + // System.out.println("Pub complete" + new + // String(token.getMessage().getPayload())); + System.out.println("发送了一条数据"); + } + + /** + * 在接收到消息时调用messageArrived这个回调。 + * 一个订阅的主题。 + */ + @Override + public void messageArrived(String topic, MqttMessage message) + throws Exception { + + Map headers = new HashMap(); + headers.put("id", "123");//数据包带有唯一标示 + headers.put("time", String.valueOf(System.currentTimeMillis()));//设备的时间戳 + Event flumeEvent = EventBuilder.withBody(message.getPayload(), + headers); + try { + getChannelProcessor().processEvent(flumeEvent); + System.out.println("消息到达-------->"); + } catch (Exception e) { + // TODO: handle exception + e.printStackTrace(); + } + + } + + /** + * + * runClient The main functionality of this simple example. Create a + * MQTT client, connect to broker, pub/sub, disconnect. + * + */ + public void runClient() { + // setup MQTT Client + String clientID = M2MIO_THING; + connOpt = new MqttConnectOptions(); + + connOpt.setCleanSession(true); + connOpt.setKeepAliveInterval(3000); + connOpt.setUserName(M2MIO_USERNAME); + connOpt.setPassword(M2MIO_PASSWORD_MD5.toCharArray()); + + // Connect to Broker + try { + myClient = new MqttClient(BROKER_URL, clientID); + myClient.setCallback(this); + myClient.connect(connOpt); + } catch (MqttException e) { + e.printStackTrace(); + System.exit(-1); + } + + System.out.println("Connected to " + BROKER_URL); + + // setup topic + // topics on m2m.io are in the form // + /** + * 方式一 + */ + /*String myTopic = M2MIO_DOMAIN + "/" + M2MIO_STUFF + "/" + + M2MIO_THING; + System.out.println("myTopic:" + myTopic); + MqttTopic topic = myClient.getTopic(myTopic);*/ + /** + * 方式二 + */ + MqttTopic topic = myClient.getTopic(M2MIO_THING); + + // subscribe to topic if subscriber + if (subscriber) { + try { + int subQoS = 2; + // myClient.subscribe(myTopic, subQoS); + myClient.subscribe(M2MIO_THING, subQoS); + + } catch (Exception e) { + e.printStackTrace(); + } + } + + // publish messages if publisher + if (publisher) { + for (int i = 1; i <= 10; i++) { + String pubMsg = "{\"pubmsg\":" + i + "}"; + int pubQoS = 2; + MqttMessage message = new MqttMessage(pubMsg.getBytes()); + message.setQos(pubQoS); + message.setRetained(false); + + // Publish the message + System.out.println("Publishing to topic \"" + topic + + "\" qos " + pubQoS); + MqttDeliveryToken token = null; + try { + // publish message to broker + token = topic.publish(message); + // Wait until the message has been delivered to the + // broker + token.waitForCompletion(); + Thread.sleep(2000); + } catch (Exception e) { + e.printStackTrace(); + } + } + } + + // disconnect + try { + // wait to ensure subscribed messages are delivered + if (subscriber) { + while (true) { + Thread.sleep(5000); + } + } + // myClient.disconnect(); + } catch (Exception e) { + e.printStackTrace(); + } finally { + } + } + + } + +} diff --git a/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource2.java b/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource2.java new file mode 100644 index 0000000..111d8a7 --- /dev/null +++ b/Big_data_example/flume/src/main/java/com/atguigu/flume/test/MQTTSource2.java @@ -0,0 +1,232 @@ +package com.atguigu.flume.test; + + +import java.util.HashMap; +import java.util.Map; + +import org.apache.flume.Context; +import org.apache.flume.Event; +import org.apache.flume.EventDrivenSource; +import org.apache.flume.conf.Configurable; +import org.apache.flume.event.EventBuilder; +import org.apache.flume.source.AbstractSource; +import org.eclipse.paho.client.mqttv3.IMqttDeliveryToken; +import org.eclipse.paho.client.mqttv3.MqttCallback; +import org.eclipse.paho.client.mqttv3.MqttClient; +import org.eclipse.paho.client.mqttv3.MqttConnectOptions; +import org.eclipse.paho.client.mqttv3.MqttDeliveryToken; +import org.eclipse.paho.client.mqttv3.MqttException; +import org.eclipse.paho.client.mqttv3.MqttMessage; +import org.eclipse.paho.client.mqttv3.MqttTopic; + +public class MQTTSource2 extends AbstractSource implements EventDrivenSource, + Configurable { + /** + * The initialization method for the Source. The context contains all the + * Flume configuration info, and can be used to retrieve any configuration + * values necessary to set up the Source. + */ + @Override + public void configure(Context arg0) { + // TODO Auto-generated method stub + + } + + SimpleMqttClient client = null; + + /** + * Start any dependent systems and begin processing events. + */ + @Override + public void start() { + // TODO Auto-generated method stub + // super.start(); + client = new SimpleMqttClient(); + client.runClient(); + } + + /** + * Stop processing events and shut any dependent systems down. + */ + @Override + public void stop() { + // TODO Auto-generated method stub + // super.stop(); + if (client != null) { + client.closeConn(); + } + } + + // public static void main(String[] args) { + // SimpleMqttClient smc = new SimpleMqttClient(); + // smc.runClient(); + // } + + public class SimpleMqttClient implements MqttCallback { + + MqttClient myClient; + MqttConnectOptions connOpt; + + String BROKER_URL = "tcp://192.168.116.202:1883"; + String M2MIO_DOMAIN = "192.168.116.202"; + String M2MIO_STUFF = "yhx"; + String M2MIO_THING = "yhx_flume"; + // String M2MIO_USERNAME = ""; + // String M2MIO_PASSWORD_MD5 = + // ""; + + Boolean subscriber = true; + Boolean publisher = false; + + /** + * connectionLost This callback is invoked upon losing the MQTT + * connection. + */ + @Override + public void connectionLost(Throwable t) { + System.out.println("Connection lost!"); + // code to reconnect to the broker would go here if desired + } + + public void closeConn() { + if (myClient != null) { + if (myClient.isConnected()) { + try { + myClient.disconnect(); + } catch (MqttException e) { + // TODO Auto-generated catch block + e.printStackTrace(); + } + } + } + } + + /** + * deliveryComplete This callback is invoked when a message published by + * this client is successfully received by the broker. + */ + @Override + public void deliveryComplete(IMqttDeliveryToken token) { + // System.out.println("Pub complete" + new + // String(token.getMessage().getPayload())); + } + + /** + * messageArrived This callback is invoked when a message is received on + * a subscribed topic. + */ + @Override + public void messageArrived(String topic, MqttMessage message) + throws Exception { + // System.out + // .println("-------------------------------------------------"); + // // System.out.println("| Topic:" + topic.getName()); + // System.out.println("| Topic:" + topic); + // System.out + // .println("| Message: " + new String(message.getPayload())); + // System.out + // .println("-------------------------------------------------"); + + + Map headers = new HashMap(); + //headers.put("curDate", df.format(new Date())); + + Event flumeEvent = EventBuilder.withBody(message.getPayload(), + headers); + try { + getChannelProcessor().processEvent(flumeEvent); + } catch (Exception e) { + // TODO: handle exception + e.printStackTrace(); + } + + } + + /** + * runClient The main functionality of this simple example. Create a + * MQTT client, connect to broker, pub/sub, disconnect. + */ + public void runClient() { + // setup MQTT Client + String clientID = M2MIO_THING; + connOpt = new MqttConnectOptions(); + + connOpt.setCleanSession(true); + connOpt.setKeepAliveInterval(3000); + // connOpt.setUserName(M2MIO_USERNAME); + // connOpt.setPassword(M2MIO_PASSWORD_MD5.toCharArray()); + + // Connect to Broker + try { + myClient = new MqttClient(BROKER_URL, clientID); + myClient.setCallback(this); + myClient.connect(connOpt); + } catch (MqttException e) { + e.printStackTrace(); + System.exit(-1); + } + + System.out.println("Connected to " + BROKER_URL); + + // setup topic + // topics on m2m.io are in the form // + String myTopic = M2MIO_DOMAIN + "/" + M2MIO_STUFF + "/" + + M2MIO_THING; + System.out.println("myTopic:" + myTopic); + MqttTopic topic = myClient.getTopic(myTopic); + + // subscribe to topic if subscriber + if (subscriber) { + try { + int subQoS = 0; + myClient.subscribe(myTopic, subQoS); + + } catch (Exception e) { + e.printStackTrace(); + } + } + + // publish messages if publisher + if (publisher) { + for (int i = 1; i <= 10; i++) { + String pubMsg = "{\"pubmsg\":" + i + "}"; + int pubQoS = 0; + MqttMessage message = new MqttMessage(pubMsg.getBytes()); + message.setQos(pubQoS); + message.setRetained(false); + + // Publish the message + System.out.println("Publishing to topic \"" + topic + + "\" qos " + pubQoS); + MqttDeliveryToken token = null; + try { + // publish message to broker + token = topic.publish(message); + // Wait until the message has been delivered to the + // broker + token.waitForCompletion(); + Thread.sleep(100); + } catch (Exception e) { + e.printStackTrace(); + } + } + } + + // disconnect + try { + // wait to ensure subscribed messages are delivered + if (subscriber) { + while (true) { + Thread.sleep(5000); + } + } + // myClient.disconnect(); + } catch (Exception e) { + e.printStackTrace(); + } finally { + } + } + + } + +} diff --git a/Big_data_example/log/agent.log b/Big_data_example/log/agent.log index 64a4d54..5e8c2fe 100644 --- a/Big_data_example/log/agent.log +++ b/Big_data_example/log/agent.log @@ -687,3 +687,4357 @@ Caused by: java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.Lz at org.apache.hadoop.io.compress.CompressionCodecFactory.getCodecClasses(CompressionCodecFactory.java:132) ... 115 more 2022-05-13 15:34:00,710 ERROR --- [ main] org.apache.hadoop.hdfs.KeyProviderCache (line: 87) : Could not find uri with key [dfs.encryption.key.provider.uri] to create a keyProvider !! +2022-05-20 12:03:52,360 ERROR --- [ main] org.apache.spark.ml.util.Instrumentation (line: 73) : java.lang.UnsupportedOperationException: empty collection + at org.apache.spark.rdd.RDD.$anonfun$treeReduce$6(RDD.scala:1128) + at scala.Option.getOrElse(Option.scala:189) + at org.apache.spark.rdd.RDD.$anonfun$treeReduce$1(RDD.scala:1128) + at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) + at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) + at org.apache.spark.rdd.RDD.withScope(RDD.scala:388) + at org.apache.spark.rdd.RDD.treeReduce(RDD.scala:1105) + at org.apache.spark.ml.feature.VectorIndexer.fit(VectorIndexer.scala:152) + at org.apache.spark.ml.feature.VectorIndexer.fit(VectorIndexer.scala:119) + at org.apache.spark.ml.Pipeline.$anonfun$fit$5(Pipeline.scala:151) + at org.apache.spark.ml.MLEvents.withFitEvent(events.scala:132) + at org.apache.spark.ml.MLEvents.withFitEvent$(events.scala:125) + at org.apache.spark.ml.util.Instrumentation.withFitEvent(Instrumentation.scala:42) + at org.apache.spark.ml.Pipeline.$anonfun$fit$4(Pipeline.scala:151) + at scala.collection.Iterator.foreach(Iterator.scala:943) + at scala.collection.Iterator.foreach$(Iterator.scala:943) + at scala.collection.AbstractIterator.foreach(Iterator.scala:1431) + at scala.collection.IterableViewLike$Transformed.foreach(IterableViewLike.scala:47) + at scala.collection.IterableViewLike$Transformed.foreach$(IterableViewLike.scala:47) + at scala.collection.SeqViewLike$AbstractTransformed.foreach(SeqViewLike.scala:40) + at org.apache.spark.ml.Pipeline.$anonfun$fit$2(Pipeline.scala:147) + at org.apache.spark.ml.MLEvents.withFitEvent(events.scala:132) + at org.apache.spark.ml.MLEvents.withFitEvent$(events.scala:125) + at org.apache.spark.ml.util.Instrumentation.withFitEvent(Instrumentation.scala:42) + at org.apache.spark.ml.Pipeline.$anonfun$fit$1(Pipeline.scala:133) + at org.apache.spark.ml.util.Instrumentation$.$anonfun$instrumented$1(Instrumentation.scala:191) + at scala.util.Try$.apply(Try.scala:213) + at org.apache.spark.ml.util.Instrumentation$.instrumented(Instrumentation.scala:191) + at org.apache.spark.ml.Pipeline.fit(Pipeline.scala:133) + at com.atguigu.userprofile.ml.pipline.MyPipeLine.train(MyPipeLine.scala:147) + at com.atguigu.userprofile.ml.train.StudGenderTrain$.main(StudGenderTrain.scala:54) + at com.atguigu.userprofile.ml.train.StudGenderTrain.main(StudGenderTrain.scala) + +2022-05-20 12:05:05,000 ERROR --- [ main] org.apache.spark.ml.util.Instrumentation (line: 73) : java.lang.UnsupportedOperationException: empty collection + at org.apache.spark.rdd.RDD.$anonfun$treeReduce$6(RDD.scala:1128) + at scala.Option.getOrElse(Option.scala:189) + at org.apache.spark.rdd.RDD.$anonfun$treeReduce$1(RDD.scala:1128) + at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) + at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) + at org.apache.spark.rdd.RDD.withScope(RDD.scala:388) + at org.apache.spark.rdd.RDD.treeReduce(RDD.scala:1105) + at org.apache.spark.ml.feature.VectorIndexer.fit(VectorIndexer.scala:152) + at org.apache.spark.ml.feature.VectorIndexer.fit(VectorIndexer.scala:119) + at org.apache.spark.ml.Pipeline.$anonfun$fit$5(Pipeline.scala:151) + at org.apache.spark.ml.MLEvents.withFitEvent(events.scala:132) + at org.apache.spark.ml.MLEvents.withFitEvent$(events.scala:125) + at org.apache.spark.ml.util.Instrumentation.withFitEvent(Instrumentation.scala:42) + at org.apache.spark.ml.Pipeline.$anonfun$fit$4(Pipeline.scala:151) + at scala.collection.Iterator.foreach(Iterator.scala:943) + at scala.collection.Iterator.foreach$(Iterator.scala:943) + at scala.collection.AbstractIterator.foreach(Iterator.scala:1431) + at scala.collection.IterableViewLike$Transformed.foreach(IterableViewLike.scala:47) + at scala.collection.IterableViewLike$Transformed.foreach$(IterableViewLike.scala:47) + at scala.collection.SeqViewLike$AbstractTransformed.foreach(SeqViewLike.scala:40) + at org.apache.spark.ml.Pipeline.$anonfun$fit$2(Pipeline.scala:147) + at org.apache.spark.ml.MLEvents.withFitEvent(events.scala:132) + at org.apache.spark.ml.MLEvents.withFitEvent$(events.scala:125) + at org.apache.spark.ml.util.Instrumentation.withFitEvent(Instrumentation.scala:42) + at org.apache.spark.ml.Pipeline.$anonfun$fit$1(Pipeline.scala:133) + at org.apache.spark.ml.util.Instrumentation$.$anonfun$instrumented$1(Instrumentation.scala:191) + at scala.util.Try$.apply(Try.scala:213) + at org.apache.spark.ml.util.Instrumentation$.instrumented(Instrumentation.scala:191) + at org.apache.spark.ml.Pipeline.fit(Pipeline.scala:133) + at com.atguigu.userprofile.ml.pipline.MyPipeLine.train(MyPipeLine.scala:147) + at com.atguigu.userprofile.ml.train.StudGenderTrain$.main(StudGenderTrain.scala:55) + at com.atguigu.userprofile.ml.train.StudGenderTrain.main(StudGenderTrain.scala) + +2022-05-24 13:32:29,008 ERROR --- [ Executor task launch worker for task 7425] org.apache.spark.executor.Executor (line: 94) : Exception in task 64.0 in stage 122.0 (TID 7425) +org.apache.spark.SparkException: Failed to execute user defined function(VectorIndexerModel$$Lambda$4552/338152470: (struct,values:array>) => struct,values:array>) + at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage9.processNext(Unknown Source) + at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43) + at org.apache.spark.sql.execution.WholeStageCodegenExec$$anon$1.hasNext(WholeStageCodegenExec.scala:729) + at org.apache.spark.sql.execution.SparkPlan.$anonfun$getByteArrayRdd$1(SparkPlan.scala:340) + at org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:872) + at org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:872) + at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) + at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:349) + at org.apache.spark.rdd.RDD.iterator(RDD.scala:313) + at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) + at org.apache.spark.scheduler.Task.run(Task.scala:127) + at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:444) + at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1377) + at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:447) + at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) + at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) + at java.lang.Thread.run(Thread.java:748) +Caused by: org.apache.spark.SparkException: VectorIndexer encountered invalid value 2.0 on feature index 2. To handle or skip invalid value, try setting VectorIndexer.handleInvalid. + at org.apache.spark.ml.feature.VectorIndexerModel.$anonfun$transformFunc$3(VectorIndexer.scala:370) + at org.apache.spark.ml.feature.VectorIndexerModel.$anonfun$transformFunc$3$adapted(VectorIndexer.scala:361) + at scala.collection.immutable.Map$Map3.foreach(Map.scala:267) + at org.apache.spark.ml.feature.VectorIndexerModel.$anonfun$transformFunc$1(VectorIndexer.scala:361) + at org.apache.spark.ml.feature.VectorIndexerModel.$anonfun$transform$1(VectorIndexer.scala:429) + ... 17 more +2022-05-24 13:32:29,254 ERROR --- [ task-result-getter-2] org.apache.spark.scheduler.TaskSetManager (line: 73) : Task 64 in stage 122.0 failed 1 times; aborting job +2022-05-24 13:56:06,664 ERROR --- [ Executor task launch worker for task 8914] org.apache.spark.executor.Executor (line: 94) : Exception in task 53.0 in stage 144.0 (TID 8914) +org.apache.spark.SparkException: Failed to execute user defined function(VectorIndexerModel$$Lambda$4629/826045717: (struct,values:array>) => struct,values:array>) + at org.apache.spark.sql.catalyst.expressions.GeneratedClass$GeneratedIteratorForCodegenStage9.processNext(Unknown Source) + at org.apache.spark.sql.execution.BufferedRowIterator.hasNext(BufferedRowIterator.java:43) + at org.apache.spark.sql.execution.WholeStageCodegenExec$$anon$1.hasNext(WholeStageCodegenExec.scala:729) + at org.apache.spark.sql.execution.SparkPlan.$anonfun$getByteArrayRdd$1(SparkPlan.scala:340) + at org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:872) + at org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:872) + at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) + at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:349) + at org.apache.spark.rdd.RDD.iterator(RDD.scala:313) + at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) + at org.apache.spark.scheduler.Task.run(Task.scala:127) + at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:444) + at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1377) + at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:447) + at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) + at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) + at java.lang.Thread.run(Thread.java:748) +Caused by: org.apache.spark.SparkException: VectorIndexer encountered invalid value 14.0 on feature index 2. To handle or skip invalid value, try setting VectorIndexer.handleInvalid. + at org.apache.spark.ml.feature.VectorIndexerModel.$anonfun$transformFunc$3(VectorIndexer.scala:370) + at org.apache.spark.ml.feature.VectorIndexerModel.$anonfun$transformFunc$3$adapted(VectorIndexer.scala:361) + at scala.collection.immutable.Map$Map3.foreach(Map.scala:267) + at org.apache.spark.ml.feature.VectorIndexerModel.$anonfun$transformFunc$1(VectorIndexer.scala:361) + at org.apache.spark.ml.feature.VectorIndexerModel.$anonfun$transform$1(VectorIndexer.scala:429) + ... 17 more +2022-05-24 13:56:06,945 ERROR --- [ task-result-getter-1] org.apache.spark.scheduler.TaskSetManager (line: 73) : Task 53 in stage 144.0 failed 1 times; aborting job +2022-05-24 19:06:24,812 ERROR --- [ main] DataNucleus.Datastore.Schema (line: 125) : Failed initialising database. +Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + +org.datanucleus.exceptions.NucleusDataStoreException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:498) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + ... 108 more +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +Nested Throwables StackTrace: +java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +2022-05-24 19:06:24,842 ERROR --- [ main] DataNucleus.Datastore (line: 125) : Exception thrown creating StoreManager. See the nested exception +Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + +org.datanucleus.exceptions.NucleusDataStoreException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:498) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + ... 108 more +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +Nested Throwables StackTrace: +java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:655) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +2022-05-24 19:07:07,524 ERROR --- [ main] DataNucleus.Datastore.Schema (line: 125) : Failed initialising database. +Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + +org.datanucleus.exceptions.NucleusDataStoreException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:498) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + ... 108 more +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +Nested Throwables StackTrace: +java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +2022-05-24 19:07:07,585 ERROR --- [ main] DataNucleus.Datastore (line: 125) : Exception thrown creating StoreManager. See the nested exception +Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + +org.datanucleus.exceptions.NucleusDataStoreException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:498) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + ... 108 more +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +Nested Throwables StackTrace: +java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +2022-05-24 19:07:07,629 ERROR --- [ main] org.apache.hadoop.hive.metastore.RetryingHMSHandler (line: 218) : Retrying HMSHandler after 2000 ms (attempt 1 of 10) with error: javax.jdo.JDOFatalDataStoreException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at org.datanucleus.api.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:529) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:830) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +NestedThrowablesStackTrace: +java.sql.SQLException: Unable to open a test connection to the given database. JDBC url = jdbc:mysql://Ding202:3306/metastore?createDatabaseIfNotExist=true&characterEncoding=utf-8&useSSL=false, username = root. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------ +com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more +------ + + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.jolbox.bonecp.PoolUtil.generateSQLException(PoolUtil.java:192) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:422) + at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120) + at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:483) + at org.datanucleus.store.rdbms.RDBMSStoreManager.(RDBMSStoreManager.java:297) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:606) + at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301) + at org.datanucleus.NucleusContextHelper.createStoreManagerForProperties(NucleusContextHelper.java:133) + at org.datanucleus.PersistenceNucleusContextImpl.initialise(PersistenceNucleusContextImpl.java:422) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:817) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:334) + at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:213) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965) + at java.security.AccessController.doPrivileged(Native Method) + at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960) + at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808) + at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701) + at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:521) + at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:550) + at org.apache.hadoop.hive.metastore.ObjectStore.initializeHelper(ObjectStore.java:405) + at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:342) + at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:303) + at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76) + at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136) + at org.apache.hadoop.hive.metastore.RawStoreProxy.(RawStoreProxy.java:58) + at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:67) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStoreForConf(HiveMetaStore.java:628) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMSForConf(HiveMetaStore.java:594) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:588) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:659) + at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:431) + at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) + at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) + at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) + at java.lang.reflect.Method.invoke(Method.java:498) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.(RetryingHMSHandler.java:79) + at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:92) + at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6902) + at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.(HiveMetaStoreClient.java:164) + at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.(SessionHiveMetaStoreClient.java:70) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1707) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) + at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) + at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3600) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3652) + at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3632) + at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3894) + at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:248) + at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:231) + at org.apache.hadoop.hive.ql.metadata.Hive.(Hive.java:388) + at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:332) + at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:312) + at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:288) + at org.apache.spark.sql.hive.client.HiveClientImpl.client(HiveClientImpl.scala:260) + at org.apache.spark.sql.hive.client.HiveClientImpl.$anonfun$withHiveState$1(HiveClientImpl.scala:286) + at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227) + at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226) + at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:276) + at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:389) + at org.apache.spark.sql.hive.HiveExternalCatalog.$anonfun$databaseExists$1(HiveExternalCatalog.scala:221) + at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:23) + at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99) + at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:221) + at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:137) + at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:127) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:157) + at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:155) + at org.apache.spark.sql.hive.HiveSessionStateBuilder.$anonfun$catalog$2(HiveSessionStateBuilder.scala:59) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:93) + at org.apache.spark.sql.catalyst.catalog.SessionCatalog.setCurrentDatabase(SessionCatalog.scala:260) + at org.apache.spark.sql.connector.catalog.CatalogManager.setCurrentNamespace(CatalogManager.scala:113) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.$anonfun$run$2$adapted(SetCatalogAndNamespaceExec.scala:36) + at scala.Option.map(Option.scala:230) + at org.apache.spark.sql.execution.datasources.v2.SetCatalogAndNamespaceExec.run(SetCatalogAndNamespaceExec.scala:36) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result$lzycompute(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.result(V2CommandExec.scala:39) + at org.apache.spark.sql.execution.datasources.v2.V2CommandExec.executeCollect(V2CommandExec.scala:45) + at org.apache.spark.sql.Dataset.$anonfun$logicalPlan$1(Dataset.scala:229) + at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:3616) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100) + at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160) + at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) + at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3614) + at org.apache.spark.sql.Dataset.(Dataset.scala:229) + at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) + at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:606) + at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:763) + at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:601) + at com.atguigu.userprofile.ml.app.BusiGenderApp$.main(BusiGenderApp.scala:69) + at com.atguigu.userprofile.ml.app.BusiGenderApp.main(BusiGenderApp.scala) +Caused by: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure + +The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server. + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:990) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:342) + at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2197) + at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2230) + at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2025) + at com.mysql.jdbc.ConnectionImpl.(ConnectionImpl.java:778) + at com.mysql.jdbc.JDBC4Connection.(JDBC4Connection.java:47) + at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) + at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) + at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) + at java.lang.reflect.Constructor.newInstance(Constructor.java:423) + at com.mysql.jdbc.Util.handleNewInstance(Util.java:425) + at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:386) + at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:330) + at java.sql.DriverManager.getConnection(DriverManager.java:664) + at java.sql.DriverManager.getConnection(DriverManager.java:208) + at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361) + at com.jolbox.bonecp.BoneCP.(BoneCP.java:416) + ... 110 more +Caused by: java.net.ConnectException: Connection timed out: connect + at java.net.DualStackPlainSocketImpl.connect0(Native Method) + at java.net.DualStackPlainSocketImpl.socketConnect(DualStackPlainSocketImpl.java:75) + at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:476) + at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:218) + at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:200) + at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:162) + at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:394) + at java.net.Socket.connect(Socket.java:606) + at com.mysql.jdbc.StandardSocketFactory.connect(StandardSocketFactory.java:211) + at com.mysql.jdbc.MysqlIO.(MysqlIO.java:301) + ... 126 more + +2022-05-24 20:14:09,927 ERROR --- [ main] org.apache.hadoop.hdfs.KeyProviderCache (line: 87) : Could not find uri with key [dfs.encryption.key.provider.uri] to create a keyProvider !! diff --git a/Big_data_example/rt-gmall-parent/gmall-logger/src/main/java/com/atguigu/rtgmall/controller/FirstController.java b/Big_data_example/rt-gmall-parent/gmall-logger/src/main/java/com/atguigu/rtgmall/controller/FirstController.java new file mode 100644 index 0000000..7c63fb4 --- /dev/null +++ b/Big_data_example/rt-gmall-parent/gmall-logger/src/main/java/com/atguigu/rtgmall/controller/FirstController.java @@ -0,0 +1,20 @@ +package com.atguigu.rtgmall.controller; + +import org.springframework.stereotype.Controller; +import org.springframework.web.bind.annotation.RequestBody; +import org.springframework.web.bind.annotation.RequestMapping; +import org.springframework.web.bind.annotation.RestController; + +@RestController +public class FirstController { + + @RequestMapping("/first") + public String first(String username, String password) { + + System.out.println(username + ":::" + password); + return "success"; + + } + + +} diff --git a/Big_data_example/rt-gmall-parent/gmall-logger/src/main/java/com/atguigu/rtgmall/controller/LoggerController.java b/Big_data_example/rt-gmall-parent/gmall-logger/src/main/java/com/atguigu/rtgmall/controller/LoggerController.java new file mode 100644 index 0000000..de6927e --- /dev/null +++ b/Big_data_example/rt-gmall-parent/gmall-logger/src/main/java/com/atguigu/rtgmall/controller/LoggerController.java @@ -0,0 +1,55 @@ +package com.atguigu.rtgmall.controller; + + +import lombok.extern.slf4j.Slf4j; +import org.apache.kafka.clients.producer.KafkaProducer; +import org.apache.kafka.clients.producer.ProducerConfig; +import org.apache.kafka.clients.producer.ProducerRecord; +import org.springframework.beans.factory.annotation.Autowired; +import org.springframework.kafka.core.KafkaTemplate; +import org.springframework.web.bind.annotation.RequestMapping; +import org.springframework.web.bind.annotation.RequestParam; +import org.springframework.web.bind.annotation.RestController; + +import java.util.Properties; + +/** + * 日志数据的采集 + */ +@RestController +@Slf4j +public class LoggerController { + + //log对象也可以自己生成,反之则用slf4j注解实现 +// private static final org.slf4j.Logger log = org.slf4j.LoggerFactory.getLogger(LogExample.class); + + @Autowired + private KafkaTemplate kafkaTemplate; + + + @RequestMapping("/applog") + public String log(@RequestParam("param")String logStr){ + //1.打印输出到控制台 +// System.out.println(logStr); + //2.落盘 ->具体查看resource下面的logback.xml文件 + log.info(logStr); + //3.将日志发送到kafka主题 +// Properties properties = new Properties(); +// properties.setProperty(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG,"Ding202:9092"); +// properties.setProperty(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG,""); +// properties.setProperty(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG,""); +// KafkaProducer kafkaProducer = new KafkaProducer(properties); +// kafkaProducer.send( +// new ProducerRecord("ods_base_log",logStr) +// ); + //spring内置的kafka方式 ->这里默认是黏性分区 + kafkaTemplate.send("ods_base_log",logStr); + + + return "success"; + } + + + + +} diff --git a/Big_data_example/rt-gmall-parent/gmall-logger/src/main/resources/logback.xml b/Big_data_example/rt-gmall-parent/gmall-logger/src/main/resources/logback.xml new file mode 100644 index 0000000..e997464 --- /dev/null +++ b/Big_data_example/rt-gmall-parent/gmall-logger/src/main/resources/logback.xml @@ -0,0 +1,32 @@ + + + + + + + %msg%n + + + + + ${LOG_HOME}/app.log + + ${LOG_HOME}/app.%d{yyyy-MM-dd}.log + + + %msg%n + + + + + + + + + + + + + + diff --git a/Big_data_example/rt-gmall-parent/gmall-realtime/pom.xml b/Big_data_example/rt-gmall-parent/gmall-realtime/pom.xml new file mode 100644 index 0000000..1e98a7b --- /dev/null +++ b/Big_data_example/rt-gmall-parent/gmall-realtime/pom.xml @@ -0,0 +1,118 @@ + + + + rt-gmall-parent + com.atguigu.rtgmall + 1.0-SNAPSHOT + + 4.0.0 + + gmall-realtime + + + 1.8 + ${java.version} + ${java.version} + 1.12.0 + 2.12 + 3.1.3 + + + + + org.apache.flink + flink-java + ${flink.version} + + + + org.apache.flink + flink-streaming-java_${scala.version} + ${flink.version} + + + + org.apache.flink + flink-connector-kafka_${scala.version} + ${flink.version} + + + + org.apache.flink + flink-clients_${scala.version} + ${flink.version} + + + + org.apache.flink + flink-cep_${scala.version} + ${flink.version} + + + + org.apache.flink + flink-json + ${flink.version} + + + + com.alibaba + fastjson + 1.2.68 + + + + org.apache.hadoop + hadoop-client + ${hadoop.version} + + + + org.slf4j + slf4j-api + 1.7.25 + + + + org.slf4j + slf4j-log4j12 + 1.7.25 + + + + org.apache.logging.log4j + log4j-to-slf4j + 2.14.0 + + + + + + + + org.apache.maven.plugins + maven-assembly-plugin + 3.0.0 + + + jar-with-dependencies + + + + + make-assembly + package + + single + + + + + + + + + + \ No newline at end of file diff --git a/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/java/com/atguigu/gmall/realtime/app/dwd/BaseLogApp.java b/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/java/com/atguigu/gmall/realtime/app/dwd/BaseLogApp.java new file mode 100644 index 0000000..a38980c --- /dev/null +++ b/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/java/com/atguigu/gmall/realtime/app/dwd/BaseLogApp.java @@ -0,0 +1,110 @@ +package com.atguigu.gmall.realtime.app.dwd; + +import com.alibaba.fastjson.JSON; +import com.alibaba.fastjson.JSONObject; +import com.atguigu.gmall.realtime.utils.MyKafkaUtils; +import org.apache.flink.api.common.functions.MapFunction; +import org.apache.flink.api.common.restartstrategy.RestartStrategies; +import org.apache.flink.api.common.serialization.SimpleStringSchema; +import org.apache.flink.runtime.state.filesystem.FsStateBackend; +import org.apache.flink.streaming.api.CheckpointingMode; +import org.apache.flink.streaming.api.datastream.DataStreamSource; +import org.apache.flink.streaming.api.datastream.SingleOutputStreamOperator; +import org.apache.flink.streaming.api.environment.CheckpointConfig; +import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment; +import org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer; +import org.apache.kafka.clients.consumer.ConsumerConfig; + +import java.util.Properties; + +/** + * 对日志数据进行分流操作 + * 启动日志、曝光日志、页面日志 + * 启动日志放到启动侧输出流中 + * 曝光日志放到曝光侧输出流中 + * 页面日志放到页面侧输出流中 + * 将不同流的数据协会kafka的dwd主题中 + */ +public class BaseLogApp { + + public static void main(String[] args) throws Exception { + + //TODO 1.基本环境准备 + //1.1 流处理环境 + StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); + + //1.2 设置并行度 + env.setParallelism(4); //这里需要与kafka对应的分区相对应 + + + + //TODO 2.检查点相关设置 + //2.1 开启检查点 + env.enableCheckpointing(5000L, CheckpointingMode.EXACTLY_ONCE); + //2.1 设置检查点超时时间 + env.getCheckpointConfig().setCheckpointTimeout(60000); + //2.1 设置重启策略 这里是固定次数的重启 ->3秒钟重启一次,一共最多重启3次 + env.setRestartStrategy(RestartStrategies.fixedDelayRestart(3,3000L)); + //2.1 设置job取消后,检查点是否保留 这里设置保留检查点-> + env.getCheckpointConfig().enableExternalizedCheckpoints(CheckpointConfig.ExternalizedCheckpointCleanup.RETAIN_ON_CANCELLATION); + //2.1 设置状态后端 即设置检查点的存储位置 内存|文件系统|RocksDB RocksDB类似于Redis会在内存中存一份,也会进行落盘 + env.setStateBackend(new FsStateBackend("hdfs://Ding202:8020/rt_gmall/gmall")); + //2 .6 指定操作HDFS的用户 + System.setProperty("HADOOP_USER_NAME","dingjiawen"); + + //TODO 3.从kafka中读取数据 + //3.1 声明消费的主题 + String topic = "ods_base_log"; + String groupId = "base_log_app_group"; + + //3.2 获取kafka消费者 + FlinkKafkaConsumer kafkaSource = MyKafkaUtils.getKafkaSource(topic, groupId); + //3.3 读取数据 封装为流 -> flink通过把kafka的offset偏移量维护在checkpoint中,保证了kafka消费的精准一致性 + DataStreamSource kafkaDS = env.addSource(kafkaSource); + + //TODO 4.读取数据进行结构转换 jsonStr ->jsonObj + //匿名内部类方式实现 + SingleOutputStreamOperator jsonObjectDs = kafkaDS.map( + new MapFunction() { + @Override + public JSONObject map(String jsonStr) throws Exception { + return JSON.parseObject(jsonStr); + } + } + ); +// //lambda表达式实现 +// kafkaDS.map( +// (MapFunction) jsonStr -> JSON.parseObject(jsonStr) +// ); + //方法的默认实现 +// kafkaDS.map(JSON::parseObject); + + jsonObjectDs.print(">>>"); + + + + + //TODO 5.新老访客状态进行修复 + + + + //TODO 6.安装日志类型对日志进行分流 + + + //TODO 7.将不同流的数据写到kafka的dwd不同的主题中 + + + + env.execute(); + + + + + + } + + + + + +} diff --git a/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/java/com/atguigu/gmall/realtime/utils/MyKafkaUtils.java b/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/java/com/atguigu/gmall/realtime/utils/MyKafkaUtils.java new file mode 100644 index 0000000..25027b1 --- /dev/null +++ b/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/java/com/atguigu/gmall/realtime/utils/MyKafkaUtils.java @@ -0,0 +1,29 @@ +package com.atguigu.gmall.realtime.utils; + + +import org.apache.flink.api.common.serialization.SimpleStringSchema; +import org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer; +import org.apache.kafka.clients.consumer.ConsumerConfig; + +import java.util.Properties; + +/** + * 操作kafka的工具类 + */ +public class MyKafkaUtils { + private static final String KAFKA_SERVER = "Ding202:9092,Ding203:9092,Ding204:9092"; + + //获取kafka的消费者 + public static FlinkKafkaConsumer getKafkaSource(String topic,String groupId) { + Properties properties = new Properties(); + properties.setProperty(ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, KAFKA_SERVER); + properties.setProperty(ConsumerConfig.GROUP_ID_CONFIG, groupId); + + + return new FlinkKafkaConsumer(topic, new SimpleStringSchema(), properties); + + + } + + +} diff --git a/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/resources/log4j.properties b/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/resources/log4j.properties new file mode 100644 index 0000000..c706282 --- /dev/null +++ b/Big_data_example/rt-gmall-parent/gmall-realtime/src/main/resources/log4j.properties @@ -0,0 +1,5 @@ +log4j.rootLogger=warn,stdout +log4j.appender.stdout=org.apache.log4j.ConsoleAppender +log4j.appender.stdout.target=System.out +log4j.appender.stdout.layout=org.apache.log4j.PatternLayout +log4j.appender.stdout.layout.ConversionPattern=%d %p [%c] - %m%n diff --git a/guli_parent/service/service_edu/edu/error/log-error-2022-04-21.0.log b/guli_parent/service/service_edu/edu/error/log-error-2022-04-21.0.log deleted file mode 100644 index 6fdf48c..0000000 --- a/guli_parent/service/service_edu/edu/error/log-error-2022-04-21.0.log +++ /dev/null @@ -1 +0,0 @@ -2022-04-21 21:01:02.160 [http-nio-8001-exec-7] ERROR c.a.s.exceptionhandler.GlobalExceptionHandler - null diff --git a/guli_parent/service/service_edu/edu/info/log-info-2022-04-21.0.log b/guli_parent/service/service_edu/edu/info/log-info-2022-04-21.0.log deleted file mode 100644 index fde37ef..0000000 --- a/guli_parent/service/service_edu/edu/info/log-info-2022-04-21.0.log +++ /dev/null @@ -1,47 +0,0 @@ -2022-04-21 20:57:35.548 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 5196 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-21 20:57:35.553 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-21 20:57:36.508 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-21 20:57:36.510 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-21 20:57:36.533 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 7ms. Found 0 repository interfaces. -2022-04-21 20:57:36.764 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-21 20:57:36.938 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-21 20:57:36.943 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-21 20:57:36.943 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-21 20:57:36.944 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-21 20:57:37.019 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-21 20:57:37.019 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 1353 ms -2022-04-21 20:57:37.161 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-21 20:57:37.324 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-21 20:57:37.654 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-21 20:57:37.727 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-21 20:57:38.052 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-21 20:57:38.061 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-21 20:57:38.081 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-21 20:57:38.219 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-21 20:57:38.233 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-21 20:57:38.237 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 3.206 seconds (JVM running for 3.756) -2022-04-21 21:00:50.190 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 13652 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-21 21:00:50.199 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-21 21:00:52.508 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-21 21:00:52.516 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-21 21:00:52.568 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 20ms. Found 0 repository interfaces. -2022-04-21 21:00:53.129 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-21 21:00:53.618 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-21 21:00:53.629 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-21 21:00:53.630 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-21 21:00:53.630 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-21 21:00:53.729 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-21 21:00:53.730 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3274 ms -2022-04-21 21:00:54.055 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-21 21:00:54.422 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-21 21:00:56.417 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-21 21:00:56.889 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-21 21:00:57.741 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-21 21:00:57.754 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-21 21:00:57.793 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-21 21:00:58.213 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-21 21:00:58.241 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-21 21:00:58.252 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 9.336 seconds (JVM running for 10.483) -2022-04-21 21:00:58.352 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-21 21:00:58.353 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-21 21:00:58.359 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 6 ms diff --git a/guli_parent/service/service_edu/edu/info/log-info-2022-04-22.0.log b/guli_parent/service/service_edu/edu/info/log-info-2022-04-22.0.log deleted file mode 100644 index 946dacb..0000000 --- a/guli_parent/service/service_edu/edu/info/log-info-2022-04-22.0.log +++ /dev/null @@ -1,200 +0,0 @@ -2022-04-22 16:11:06.739 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 20396 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-22 16:11:06.837 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-22 16:11:13.525 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-22 16:11:13.537 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-22 16:11:13.685 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 58ms. Found 0 repository interfaces. -2022-04-22 16:11:15.522 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-22 16:11:16.719 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-22 16:11:16.754 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-22 16:11:16.756 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-22 16:11:16.757 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-22 16:11:17.087 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-22 16:11:17.087 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 9841 ms -2022-04-22 16:11:17.840 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-22 16:11:18.575 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-22 16:11:21.470 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-22 16:11:22.045 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-22 16:11:25.314 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-22 16:11:25.428 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-22 16:11:25.719 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-22 16:11:27.234 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-22 16:11:27.362 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-22 16:11:27.381 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 24.506 seconds (JVM running for 29.474) -2022-04-22 16:14:26.591 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-22 16:14:26.593 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-22 16:14:26.628 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 34 ms -2022-04-22 16:22:20.475 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 8784 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-22 16:22:20.550 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-22 16:22:27.173 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-22 16:22:27.206 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-22 16:22:27.436 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 121ms. Found 0 repository interfaces. -2022-04-22 16:22:28.692 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-22 16:22:29.427 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-22 16:22:29.459 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-22 16:22:29.462 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-22 16:22:29.465 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-22 16:22:29.780 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-22 16:22:29.780 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 8769 ms -2022-04-22 16:22:30.430 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-22 16:22:30.973 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-22 16:22:32.463 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-22 16:22:32.862 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-22 16:22:34.567 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-22 16:22:34.633 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-22 16:22:34.741 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-22 16:22:35.434 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-22 16:22:35.515 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-22 16:22:35.523 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 18.867 seconds (JVM running for 24.372) -2022-04-22 16:22:41.283 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-22 16:22:41.284 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-22 16:22:41.308 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 24 ms -2022-04-22 17:22:51.944 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 15372 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-22 17:22:51.974 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-22 17:22:58.495 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-22 17:22:58.520 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-22 17:22:58.628 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 35ms. Found 0 repository interfaces. -2022-04-22 17:22:59.987 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-22 17:23:01.087 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-22 17:23:01.145 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-22 17:23:01.153 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-22 17:23:01.157 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-22 17:23:01.519 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-22 17:23:01.520 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 9275 ms -2022-04-22 17:23:02.694 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-22 17:23:03.530 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-22 17:23:06.183 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-22 17:23:06.582 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-22 17:23:08.306 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-22 17:23:08.379 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-22 17:23:08.504 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-22 17:23:09.102 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-22 17:23:09.216 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-22 17:23:09.229 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 21.399 seconds (JVM running for 29.843) -2022-04-22 17:23:20.928 [http-nio-8001-exec-2] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-22 17:23:20.932 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-22 17:23:20.957 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 22 ms -2022-04-22 19:50:55.720 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 17936 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-22 19:50:55.766 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-22 19:51:01.696 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-22 19:51:01.708 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-22 19:51:01.795 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 28ms. Found 0 repository interfaces. -2022-04-22 19:51:03.105 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-22 19:51:04.347 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-22 19:51:04.396 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-22 19:51:04.399 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-22 19:51:04.400 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-22 19:51:04.731 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-22 19:51:04.732 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 8476 ms -2022-04-22 19:51:05.201 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-22 19:51:05.761 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-22 19:51:08.772 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-22 19:51:09.359 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-22 19:51:12.004 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-22 19:51:12.098 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-22 19:51:12.325 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-22 19:51:13.449 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-22 19:51:13.553 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-22 19:51:13.572 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 20.848 seconds (JVM running for 25.253) -2022-04-22 19:51:45.842 [http-nio-8001-exec-2] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-22 19:51:45.844 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-22 19:51:45.870 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 26 ms -2022-04-22 21:02:14.650 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 19980 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-22 21:02:14.668 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-22 21:02:21.283 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-22 21:02:21.294 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-22 21:02:21.516 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 105ms. Found 0 repository interfaces. -2022-04-22 21:02:23.671 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-22 21:02:25.184 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-22 21:02:25.241 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-22 21:02:25.245 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-22 21:02:25.245 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-22 21:02:25.610 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-22 21:02:25.611 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 10704 ms -2022-04-22 21:02:26.969 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-22 21:02:27.766 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-22 21:02:30.335 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-22 21:02:30.884 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-22 21:02:32.685 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-22 21:02:32.733 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-22 21:02:32.879 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-22 21:02:33.548 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-22 21:02:33.611 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-22 21:02:33.618 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 21.724 seconds (JVM running for 26.735) -2022-04-22 21:03:27.873 [http-nio-8001-exec-3] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-22 21:03:27.878 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-22 21:03:27.894 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 15 ms -2022-04-22 21:09:29.287 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 11768 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-22 21:09:29.295 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-22 21:09:35.789 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-22 21:09:35.801 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-22 21:09:35.863 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 18ms. Found 0 repository interfaces. -2022-04-22 21:09:36.783 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-22 21:09:37.727 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-22 21:09:37.750 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-22 21:09:37.751 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-22 21:09:37.752 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-22 21:09:37.997 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-22 21:09:37.997 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 8441 ms -2022-04-22 21:09:38.591 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-22 21:09:39.120 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-22 21:09:40.639 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-22 21:09:41.132 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-22 21:09:42.654 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-22 21:09:42.715 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-22 21:09:42.840 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-22 21:09:43.505 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-22 21:09:43.575 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-22 21:09:43.588 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 16.638 seconds (JVM running for 19.034) -2022-04-22 21:09:52.962 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-22 21:09:52.965 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-22 21:09:52.998 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 31 ms -2022-04-22 21:28:43.865 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 7476 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-22 21:28:43.878 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-22 21:28:47.458 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-22 21:28:47.469 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-22 21:28:47.569 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 35ms. Found 0 repository interfaces. -2022-04-22 21:28:48.765 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-22 21:28:49.714 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-22 21:28:49.754 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-22 21:28:49.758 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-22 21:28:49.759 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-22 21:28:49.990 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-22 21:28:49.996 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 5798 ms -2022-04-22 21:28:50.621 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-22 21:28:51.164 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-22 21:28:52.918 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-22 21:28:53.533 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-22 21:28:55.146 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-22 21:28:55.204 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-22 21:28:55.335 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-22 21:28:55.994 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-22 21:28:56.069 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-22 21:28:56.097 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 14.049 seconds (JVM running for 16.394) -2022-04-22 21:29:35.486 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-22 21:29:35.487 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-22 21:29:35.567 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 79 ms -2022-04-22 21:32:40.555 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 4320 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-22 21:32:40.563 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-22 21:32:44.431 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-22 21:32:44.447 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-22 21:32:44.596 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 39ms. Found 0 repository interfaces. -2022-04-22 21:32:45.933 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-22 21:32:47.021 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-22 21:32:47.040 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-22 21:32:47.044 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-22 21:32:47.047 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-22 21:32:47.328 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-22 21:32:47.329 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 6500 ms -2022-04-22 21:32:48.019 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-22 21:32:48.994 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-22 21:32:50.604 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-22 21:32:50.959 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-22 21:32:52.480 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-22 21:32:52.547 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-22 21:32:52.687 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-22 21:32:53.849 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-22 21:32:54.051 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-22 21:32:54.069 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 15.9 seconds (JVM running for 18.414) -2022-04-22 21:32:54.397 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-22 21:32:54.398 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-22 21:32:54.494 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 94 ms diff --git a/guli_parent/service/service_edu/edu/info/log-info-2022-04-24.0.log b/guli_parent/service/service_edu/edu/info/log-info-2022-04-24.0.log deleted file mode 100644 index 8a7979b..0000000 --- a/guli_parent/service/service_edu/edu/info/log-info-2022-04-24.0.log +++ /dev/null @@ -1,444 +0,0 @@ -2022-04-24 11:11:43.893 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 20472 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 11:11:43.920 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 11:11:50.817 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 11:11:50.831 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 11:11:50.983 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 40ms. Found 0 repository interfaces. -2022-04-24 11:11:52.407 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 11:11:53.612 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 11:11:53.654 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 11:11:53.658 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 11:11:53.661 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 11:11:54.226 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 11:11:54.227 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 9958 ms -2022-04-24 11:11:55.334 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 11:11:56.295 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 11:11:58.593 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 11:11:59.244 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 11:12:01.643 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 11:12:01.717 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 11:12:01.958 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 11:12:02.955 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 11:12:03.046 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 11:12:03.059 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 23.702 seconds (JVM running for 28.285) -2022-04-24 11:12:14.122 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 11:12:14.128 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 11:12:14.162 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 32 ms -2022-04-24 15:09:04.077 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 1292 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 15:09:04.097 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 15:09:06.194 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 15:09:06.202 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 15:09:06.245 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 25ms. Found 0 repository interfaces. -2022-04-24 15:09:07.347 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 15:09:07.937 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 15:09:07.951 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 15:09:07.956 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 15:09:07.958 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 15:09:08.100 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 15:09:08.103 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3697 ms -2022-04-24 15:09:08.402 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 15:09:08.844 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 15:09:09.960 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 15:09:10.481 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 15:09:12.056 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 15:09:12.134 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 15:09:12.412 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 15:09:13.680 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 15:09:13.804 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 15:09:13.820 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 11.145 seconds (JVM running for 12.285) -2022-04-24 15:10:39.846 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 15:10:39.846 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 15:10:39.857 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 11 ms -2022-04-24 17:40:58.128 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 10652 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 17:40:58.166 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 17:41:01.970 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 17:41:01.984 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 17:41:02.074 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 49ms. Found 0 repository interfaces. -2022-04-24 17:41:02.950 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 17:41:03.407 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 17:41:03.414 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 17:41:03.415 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 17:41:03.415 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 17:41:03.556 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 17:41:03.557 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 4739 ms -2022-04-24 17:41:03.879 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 17:41:04.161 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 17:41:05.031 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 17:41:05.209 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 17:41:06.066 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 17:41:06.087 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 17:41:06.153 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 17:41:06.622 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 17:41:06.674 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 17:41:06.676 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 11.775 seconds (JVM running for 14.112) -2022-04-24 17:42:36.246 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 17:42:36.246 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 17:42:36.250 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 4 ms -2022-04-24 18:14:30.676 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 13252 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 18:14:30.709 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 18:14:32.764 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 18:14:32.767 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 18:14:32.784 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 8ms. Found 0 repository interfaces. -2022-04-24 18:14:33.006 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 18:14:33.169 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 18:14:33.175 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 18:14:33.176 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 18:14:33.176 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 18:14:33.236 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 18:14:33.236 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 2205 ms -2022-04-24 18:14:33.355 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 18:14:33.468 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 18:14:33.795 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 18:14:33.862 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 18:14:34.160 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 18:14:34.170 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 18:14:34.205 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 18:14:34.329 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 18:14:34.343 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 18:14:34.345 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 5.978 seconds (JVM running for 7.868) -2022-04-24 18:15:10.211 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 18:15:10.214 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 18:15:10.257 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 42 ms -2022-04-24 18:16:23.862 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 10992 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 18:16:23.881 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 18:16:26.215 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 18:16:26.218 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 18:16:26.284 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 25ms. Found 0 repository interfaces. -2022-04-24 18:16:26.749 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 18:16:27.155 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 18:16:27.161 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 18:16:27.162 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 18:16:27.162 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 18:16:27.320 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 18:16:27.322 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3293 ms -2022-04-24 18:16:27.749 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 18:16:27.987 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 18:16:28.807 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 18:16:28.973 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 18:16:29.757 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 18:16:29.799 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 18:16:29.863 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 18:16:30.225 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 18:16:30.281 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 18:16:30.283 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 8.212 seconds (JVM running for 9.61) -2022-04-24 18:17:04.765 [http-nio-8001-exec-3] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 18:17:04.766 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 18:17:04.814 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 46 ms -2022-04-24 18:19:36.859 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 7632 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 18:19:36.877 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 18:19:39.709 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 18:19:39.723 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 18:19:39.802 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 35ms. Found 0 repository interfaces. -2022-04-24 18:19:40.476 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 18:19:40.898 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 18:19:40.927 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 18:19:40.928 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 18:19:40.928 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 18:19:41.065 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 18:19:41.065 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3906 ms -2022-04-24 18:19:41.442 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 18:19:41.746 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 18:19:42.663 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 18:19:42.865 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 18:19:43.711 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 18:19:43.741 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 18:19:43.802 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 18:19:44.249 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 18:19:44.286 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 18:19:44.293 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 9.147 seconds (JVM running for 11.554) -2022-04-24 18:21:11.262 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 18:21:11.263 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 18:21:11.285 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 22 ms -2022-04-24 18:23:32.309 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 12460 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 18:23:32.312 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 18:23:33.221 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 18:23:33.224 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 18:23:33.254 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 12ms. Found 0 repository interfaces. -2022-04-24 18:23:33.479 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 18:23:33.638 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 18:23:33.646 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 18:23:33.646 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 18:23:33.647 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 18:23:33.709 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 18:23:33.709 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 1298 ms -2022-04-24 18:23:33.835 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 18:23:33.963 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 18:23:34.291 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 18:23:34.362 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 18:23:34.654 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 18:23:34.664 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 18:23:34.688 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 18:23:34.809 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 18:23:34.823 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 18:23:34.825 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 2.928 seconds (JVM running for 3.386) -2022-04-24 18:24:23.672 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 19008 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 18:24:23.675 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 18:24:24.343 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 18:24:24.345 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 18:24:24.361 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 6ms. Found 0 repository interfaces. -2022-04-24 18:24:24.545 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 18:24:24.702 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 18:24:24.708 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 18:24:24.708 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 18:24:24.708 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 18:24:24.769 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 18:24:24.769 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 1039 ms -2022-04-24 18:24:24.897 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 18:24:25.007 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 18:24:25.348 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 18:24:25.422 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 18:24:25.726 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 18:24:25.737 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 18:24:25.760 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 18:24:25.894 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 18:24:25.950 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 18:24:25.953 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 2.791 seconds (JVM running for 3.527) -2022-04-24 18:24:41.094 [http-nio-8001-exec-3] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 18:24:41.095 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 18:24:41.101 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 6 ms -2022-04-24 18:31:38.469 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 8608 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 18:31:38.480 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 18:31:41.826 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 18:31:41.837 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 18:31:41.923 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 37ms. Found 0 repository interfaces. -2022-04-24 18:31:42.902 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 18:31:43.572 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 18:31:43.598 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 18:31:43.601 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 18:31:43.602 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 18:31:43.877 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 18:31:43.878 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 4958 ms -2022-04-24 18:31:44.470 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 18:31:45.124 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 18:31:46.753 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 18:31:47.141 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 18:31:48.402 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 18:31:48.449 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 18:31:48.544 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 18:31:49.112 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 18:31:49.213 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 18:31:49.225 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 15.499 seconds (JVM running for 20.352) -2022-04-24 18:32:30.533 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 18:32:30.538 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 18:32:30.574 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 33 ms -2022-04-24 19:38:00.372 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 4264 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 19:38:00.395 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 19:38:04.068 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 19:38:04.068 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 19:38:04.183 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 40ms. Found 0 repository interfaces. -2022-04-24 19:38:04.841 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 19:38:05.528 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 19:38:05.538 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 19:38:05.538 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 19:38:05.538 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 19:38:05.829 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 19:38:05.829 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 5091 ms -2022-04-24 19:38:06.468 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 19:38:06.958 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 19:38:08.048 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 19:38:08.318 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 19:38:09.478 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 19:38:09.498 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 19:38:09.578 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 19:38:10.008 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 19:38:10.057 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 19:38:10.057 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 12.742 seconds (JVM running for 16.533) -2022-04-24 19:40:17.457 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 16024 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 19:40:17.457 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 19:40:18.292 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 19:40:18.292 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 19:40:18.312 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 8ms. Found 0 repository interfaces. -2022-04-24 19:40:18.587 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 19:40:18.757 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 19:40:18.762 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 19:40:18.767 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 19:40:18.767 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 19:40:18.827 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 19:40:18.827 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 1289 ms -2022-04-24 19:40:18.977 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 19:40:19.107 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 19:40:19.457 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 19:40:19.532 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 19:40:19.857 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 19:40:19.865 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 19:40:19.887 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 19:40:20.007 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 19:40:20.033 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 19:40:20.036 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 3.004 seconds (JVM running for 3.605) -2022-04-24 19:40:45.866 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 19:40:45.867 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 19:40:45.873 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 6 ms -2022-04-24 20:28:10.446 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 15048 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 20:28:10.478 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 20:28:16.933 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 20:28:16.947 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 20:28:17.087 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 57ms. Found 0 repository interfaces. -2022-04-24 20:28:18.713 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 20:28:19.982 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 20:28:20.068 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 20:28:20.072 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 20:28:20.078 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 20:28:20.472 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 20:28:20.473 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 9651 ms -2022-04-24 20:28:21.337 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 20:28:21.989 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 20:28:24.298 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 20:28:24.796 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 20:28:26.741 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 20:28:26.803 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 20:28:26.962 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 20:28:28.119 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 20:28:28.230 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 20:28:28.255 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 22.08 seconds (JVM running for 25.494) -2022-04-24 20:28:46.660 [http-nio-8001-exec-2] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 20:28:46.661 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 20:28:46.668 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 7 ms -2022-04-24 20:29:32.313 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 1788 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 20:29:32.340 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 20:29:34.862 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 20:29:34.864 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 20:29:34.916 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 34ms. Found 0 repository interfaces. -2022-04-24 20:29:35.537 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 20:29:35.922 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 20:29:35.936 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 20:29:35.936 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 20:29:35.937 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 20:29:36.079 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 20:29:36.081 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3334 ms -2022-04-24 20:29:36.545 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 20:29:36.909 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 20:29:37.830 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 20:29:38.165 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 20:29:39.286 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 20:29:39.315 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 20:29:39.370 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 20:29:39.654 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 20:29:39.691 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 20:29:39.703 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 9.074 seconds (JVM running for 10.721) -2022-04-24 20:29:55.430 [http-nio-8001-exec-2] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 20:29:55.431 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 20:29:55.459 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 25 ms -2022-04-24 20:31:08.056 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 1728 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 20:31:08.074 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 20:31:11.555 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 20:31:11.563 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 20:31:11.630 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 15ms. Found 0 repository interfaces. -2022-04-24 20:31:12.355 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 20:31:12.753 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 20:31:12.764 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 20:31:12.767 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 20:31:12.769 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 20:31:12.897 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 20:31:12.899 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 4535 ms -2022-04-24 20:31:13.263 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 20:31:13.518 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 20:31:14.332 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 20:31:14.491 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 20:31:15.223 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 20:31:15.265 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 20:31:15.300 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 20:31:15.611 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 20:31:15.628 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 20:31:15.639 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 9.283 seconds (JVM running for 11.25) -2022-04-24 20:31:27.575 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 20:31:27.576 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 20:31:27.601 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 25 ms -2022-04-24 20:32:33.005 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 13060 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 20:32:33.123 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 20:32:35.259 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 20:32:35.261 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 20:32:35.319 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 48ms. Found 0 repository interfaces. -2022-04-24 20:32:35.920 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 20:32:36.295 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 20:32:36.316 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 20:32:36.317 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 20:32:36.317 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 20:32:36.450 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 20:32:36.450 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3175 ms -2022-04-24 20:32:36.871 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 20:32:37.200 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 20:32:38.069 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 20:32:38.262 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 20:32:39.053 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 20:32:39.071 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 20:32:39.140 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 20:32:39.537 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 20:32:39.592 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 20:32:39.604 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 8.293 seconds (JVM running for 9.959) -2022-04-24 20:32:44.459 [http-nio-8001-exec-3] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 20:32:44.463 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 20:32:44.495 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 30 ms -2022-04-24 20:33:57.452 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 15836 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 20:33:57.618 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 20:34:00.419 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 20:34:00.441 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 20:34:00.585 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 46ms. Found 0 repository interfaces. -2022-04-24 20:34:01.452 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 20:34:01.964 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 20:34:01.970 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 20:34:01.970 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 20:34:01.970 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 20:34:02.162 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 20:34:02.162 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 4315 ms -2022-04-24 20:34:02.664 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 20:34:03.040 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 20:34:04.181 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 20:34:04.384 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 20:34:05.122 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 20:34:05.160 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 20:34:05.206 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 20:34:05.556 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 20:34:05.579 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 20:34:05.582 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 9.748 seconds (JVM running for 11.416) -2022-04-24 20:35:05.035 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 20:35:05.036 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 20:35:05.040 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 4 ms -2022-04-24 20:37:15.506 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 4980 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 20:37:15.521 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 20:37:18.304 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 20:37:18.313 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 20:37:18.347 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 13ms. Found 0 repository interfaces. -2022-04-24 20:37:18.981 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 20:37:19.369 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 20:37:19.382 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 20:37:19.383 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 20:37:19.383 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 20:37:19.506 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 20:37:19.508 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3699 ms -2022-04-24 20:37:19.822 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 20:37:20.134 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 20:37:21.116 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 20:37:21.275 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 20:37:21.986 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 20:37:22.005 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 20:37:22.068 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 20:37:22.377 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 20:37:22.403 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 20:37:22.405 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 8.472 seconds (JVM running for 10.023) -2022-04-24 20:37:34.812 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 20:37:34.813 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 20:37:34.839 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 26 ms -2022-04-24 20:38:09.069 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 10216 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-24 20:38:09.080 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-24 20:38:11.570 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-24 20:38:11.572 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-24 20:38:11.623 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 32ms. Found 0 repository interfaces. -2022-04-24 20:38:12.325 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-24 20:38:12.719 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-24 20:38:12.744 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-24 20:38:12.746 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-24 20:38:12.748 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-24 20:38:12.874 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-24 20:38:12.874 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3531 ms -2022-04-24 20:38:13.200 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-24 20:38:13.462 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-24 20:38:14.361 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-24 20:38:14.518 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-24 20:38:15.291 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-24 20:38:15.300 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-24 20:38:15.362 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-24 20:38:15.652 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-24 20:38:15.695 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-24 20:38:15.697 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 8.379 seconds (JVM running for 10.235) -2022-04-24 20:38:19.835 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-24 20:38:19.836 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-24 20:38:19.850 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 13 ms diff --git a/guli_parent/service/service_edu/edu/info/log-info-2022-04-25.0.log b/guli_parent/service/service_edu/edu/info/log-info-2022-04-25.0.log deleted file mode 100644 index 47c7572..0000000 --- a/guli_parent/service/service_edu/edu/info/log-info-2022-04-25.0.log +++ /dev/null @@ -1,200 +0,0 @@ -2022-04-25 10:55:47.916 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 6060 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-25 10:55:47.921 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-25 10:55:50.907 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-25 10:55:50.909 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-25 10:55:50.982 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 33ms. Found 0 repository interfaces. -2022-04-25 10:55:51.396 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-25 10:55:51.874 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-25 10:55:51.906 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-25 10:55:51.906 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-25 10:55:51.906 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-25 10:55:52.048 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-25 10:55:52.049 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3866 ms -2022-04-25 10:55:52.448 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-25 10:55:52.657 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-25 10:55:53.661 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-25 10:55:53.801 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-25 10:55:54.651 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-25 10:55:54.663 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-25 10:55:54.722 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-25 10:55:55.000 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-25 10:55:55.047 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-25 10:55:55.052 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 9.125 seconds (JVM running for 11.296) -2022-04-25 10:56:29.975 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-25 10:56:29.977 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-25 10:56:29.996 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 18 ms -2022-04-25 11:55:14.432 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 8096 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-25 11:55:14.435 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-25 11:55:16.999 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-25 11:55:17.003 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-25 11:55:17.095 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 40ms. Found 0 repository interfaces. -2022-04-25 11:55:17.613 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-25 11:55:18.009 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-25 11:55:18.015 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-25 11:55:18.018 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-25 11:55:18.019 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-25 11:55:18.361 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-25 11:55:18.361 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3763 ms -2022-04-25 11:55:18.695 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-25 11:55:19.039 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-25 11:55:20.165 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-25 11:55:20.340 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-25 11:55:20.988 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-25 11:55:21.023 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-25 11:55:21.069 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-25 11:55:21.406 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-25 11:55:21.426 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-25 11:55:21.428 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 9.584 seconds (JVM running for 12.722) -2022-04-25 11:56:12.089 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-25 11:56:12.089 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-25 11:56:12.098 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 9 ms -2022-04-25 11:59:23.301 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 6836 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-25 11:59:23.323 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-25 11:59:29.258 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-25 11:59:29.259 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-25 11:59:29.283 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 14ms. Found 0 repository interfaces. -2022-04-25 11:59:29.845 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-25 11:59:30.474 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-25 11:59:30.483 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-25 11:59:30.484 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-25 11:59:30.484 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-25 11:59:30.621 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-25 11:59:30.622 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 6908 ms -2022-04-25 11:59:30.968 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-25 11:59:31.221 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-25 11:59:32.223 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-25 11:59:32.417 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-25 11:59:33.023 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-25 11:59:33.035 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-25 11:59:33.092 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-25 11:59:33.376 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-25 11:59:33.415 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-25 11:59:33.417 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 13.949 seconds (JVM running for 18.534) -2022-04-25 11:59:34.890 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-25 11:59:34.890 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-25 11:59:34.914 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 23 ms -2022-04-25 12:01:40.333 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 3804 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-25 12:01:40.403 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-25 12:01:42.726 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-25 12:01:42.728 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-25 12:01:42.798 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 11ms. Found 0 repository interfaces. -2022-04-25 12:01:43.430 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-25 12:01:43.801 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-25 12:01:43.829 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-25 12:01:43.833 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-25 12:01:43.834 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-25 12:01:43.952 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-25 12:01:43.954 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3429 ms -2022-04-25 12:01:44.307 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-25 12:01:44.536 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-25 12:01:45.603 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-25 12:01:45.799 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-25 12:01:46.514 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-25 12:01:46.530 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-25 12:01:46.582 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-25 12:01:46.865 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-25 12:01:46.910 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-25 12:01:46.912 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 8.617 seconds (JVM running for 12.716) -2022-04-25 12:01:51.334 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-25 12:01:51.334 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-25 12:01:51.346 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 12 ms -2022-04-25 12:03:38.422 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 4976 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-25 12:03:38.437 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-25 12:03:40.592 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-25 12:03:40.594 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-25 12:03:40.656 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 10ms. Found 0 repository interfaces. -2022-04-25 12:03:41.190 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-25 12:03:41.523 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-25 12:03:41.530 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-25 12:03:41.530 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-25 12:03:41.531 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-25 12:03:41.667 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-25 12:03:41.668 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 2782 ms -2022-04-25 12:03:41.970 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-25 12:03:42.194 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-25 12:03:43.194 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-25 12:03:43.386 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-25 12:03:44.194 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-25 12:03:44.221 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-25 12:03:44.270 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-25 12:03:44.810 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-25 12:03:44.891 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-25 12:03:44.904 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 8.126 seconds (JVM running for 9.781) -2022-04-25 12:03:48.246 [http-nio-8001-exec-2] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-25 12:03:48.247 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-25 12:03:48.272 [http-nio-8001-exec-2] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 24 ms -2022-04-25 12:05:36.662 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 4432 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-25 12:05:36.693 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-25 12:05:39.522 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-25 12:05:39.535 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-25 12:05:39.603 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 38ms. Found 0 repository interfaces. -2022-04-25 12:05:40.148 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-25 12:05:40.589 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-25 12:05:40.599 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-25 12:05:40.602 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-25 12:05:40.603 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-25 12:05:40.739 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-25 12:05:40.739 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3617 ms -2022-04-25 12:05:41.141 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-25 12:05:41.473 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-25 12:05:42.476 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-25 12:05:42.651 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-25 12:05:43.412 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-25 12:05:43.455 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-25 12:05:43.517 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-25 12:05:43.861 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-25 12:05:43.891 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-25 12:05:43.901 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 9.607 seconds (JVM running for 12.458) -2022-04-25 12:05:52.260 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-25 12:05:52.261 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-25 12:05:52.267 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 5 ms -2022-04-25 12:07:16.304 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 604 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-25 12:07:16.334 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-25 12:07:17.131 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-25 12:07:17.133 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-25 12:07:17.159 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 12ms. Found 0 repository interfaces. -2022-04-25 12:07:17.402 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-25 12:07:17.573 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-25 12:07:17.579 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-25 12:07:17.580 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-25 12:07:17.580 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-25 12:07:17.645 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-25 12:07:17.645 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 1284 ms -2022-04-25 12:07:17.766 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-25 12:07:17.889 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-25 12:07:18.463 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-25 12:07:18.543 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-25 12:07:18.871 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-25 12:07:18.881 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-25 12:07:18.903 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-25 12:07:19.030 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-25 12:07:19.045 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-25 12:07:19.047 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 3.142 seconds (JVM running for 3.633) -2022-04-25 12:07:20.759 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-25 12:07:20.759 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-25 12:07:20.766 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 6 ms -2022-04-25 12:12:48.641 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 2676 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-25 12:12:48.664 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-25 12:12:53.923 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-25 12:12:53.942 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-25 12:12:54.077 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 60ms. Found 0 repository interfaces. -2022-04-25 12:12:55.405 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-25 12:12:56.446 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-25 12:12:56.483 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-25 12:12:56.486 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-25 12:12:56.490 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-25 12:12:56.869 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-25 12:12:56.870 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 8023 ms -2022-04-25 12:12:57.857 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-25 12:12:58.648 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-25 12:13:01.757 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-25 12:13:02.229 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-25 12:13:04.382 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-25 12:13:04.468 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-25 12:13:04.682 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-25 12:13:06.135 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-25 12:13:06.235 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-25 12:13:06.254 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 21.55 seconds (JVM running for 25.839) -2022-04-25 12:13:08.606 [http-nio-8001-exec-3] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-25 12:13:08.606 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-25 12:13:08.611 [http-nio-8001-exec-3] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 5 ms diff --git a/guli_parent/service/service_edu/edu/info/log-info-2022-04-27.0.log b/guli_parent/service/service_edu/edu/info/log-info-2022-04-27.0.log deleted file mode 100644 index 0121a17..0000000 --- a/guli_parent/service/service_edu/edu/info/log-info-2022-04-27.0.log +++ /dev/null @@ -1,72 +0,0 @@ -2022-04-27 13:37:36.533 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 12832 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-27 13:37:36.538 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-27 13:37:38.522 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-27 13:37:38.523 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-27 13:37:38.570 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 33ms. Found 0 repository interfaces. -2022-04-27 13:37:38.990 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-27 13:37:39.391 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-27 13:37:39.400 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-27 13:37:39.400 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-27 13:37:39.400 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-27 13:37:39.536 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-27 13:37:39.536 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 2838 ms -2022-04-27 13:37:39.915 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-27 13:37:40.522 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-27 13:37:43.505 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-27 13:37:44.212 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-27 13:37:46.297 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-27 13:37:46.350 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-27 13:37:46.495 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-27 13:37:46.944 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-27 13:37:46.964 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-27 13:37:46.966 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 11.822 seconds (JVM running for 13.597) -2022-04-27 13:37:49.095 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-27 13:37:49.096 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-27 13:37:49.101 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 5 ms -2022-04-27 13:40:14.385 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 13328 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-27 13:40:14.389 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-27 13:40:17.223 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-27 13:40:17.224 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-27 13:40:17.272 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 35ms. Found 0 repository interfaces. -2022-04-27 13:40:17.623 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-27 13:40:17.969 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-27 13:40:17.980 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-27 13:40:17.983 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-27 13:40:17.986 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-27 13:40:18.175 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-27 13:40:18.175 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3603 ms -2022-04-27 13:40:18.485 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-27 13:40:18.682 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-27 13:40:19.568 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-27 13:40:19.716 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-27 13:40:20.285 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-27 13:40:20.297 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-27 13:40:20.354 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-27 13:40:20.642 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-27 13:40:20.691 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-27 13:40:20.693 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 7.445 seconds (JVM running for 8.662) -2022-04-27 13:40:30.313 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 12844 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-04-27 13:40:30.316 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-04-27 13:40:32.809 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-04-27 13:40:32.817 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-04-27 13:40:32.853 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 17ms. Found 0 repository interfaces. -2022-04-27 13:40:33.525 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-04-27 13:40:33.888 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-04-27 13:40:33.898 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-04-27 13:40:33.899 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-04-27 13:40:33.899 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-04-27 13:40:34.094 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-04-27 13:40:34.095 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3460 ms -2022-04-27 13:40:34.421 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-04-27 13:40:34.644 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-04-27 13:40:35.501 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-04-27 13:40:35.631 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-04-27 13:40:36.289 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-04-27 13:40:36.304 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-04-27 13:40:36.365 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-04-27 13:40:36.611 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-04-27 13:40:36.636 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-04-27 13:40:36.637 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 7.623 seconds (JVM running for 9.067) -2022-04-27 13:41:03.686 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-04-27 13:41:03.687 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-04-27 13:41:03.691 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 4 ms diff --git a/guli_parent/service/service_edu/edu/log_error.log b/guli_parent/service/service_edu/edu/log_error.log index 3632226..5b11df1 100644 --- a/guli_parent/service/service_edu/edu/log_error.log +++ b/guli_parent/service/service_edu/edu/log_error.log @@ -1,106 +1,14 @@ -2022-05-05 13:35:38.424 [http-nio-8001-exec-10] ERROR c.a.s.exceptionhandler.GlobalExceptionHandler - GuliException(code=20001, msg=执行了自定义异常处理...) - at com.atguigu.eduservice.controller.EduTeacherController.findAllTeacher(EduTeacherController.java:54) - at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) - at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) - at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) - at java.lang.reflect.Method.invoke(Method.java:498) - at org.springframework.web.method.support.InvocableHandlerMethod.doInvoke(InvocableHandlerMethod.java:190) - at org.springframework.web.method.support.InvocableHandlerMethod.invokeForRequest(InvocableHandlerMethod.java:138) - at org.springframework.web.servlet.mvc.method.annotation.ServletInvocableHandlerMethod.invokeAndHandle(ServletInvocableHandlerMethod.java:106) - at org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.invokeHandlerMethod(RequestMappingHandlerAdapter.java:888) - at org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.handleInternal(RequestMappingHandlerAdapter.java:793) - at org.springframework.web.servlet.mvc.method.AbstractHandlerMethodAdapter.handle(AbstractHandlerMethodAdapter.java:87) - at org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:1040) - at org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:943) - at org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:1006) - at org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:898) - at javax.servlet.http.HttpServlet.service(HttpServlet.java:634) - at org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:883) - at javax.servlet.http.HttpServlet.service(HttpServlet.java:741) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:53) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.springframework.web.filter.RequestContextFilter.doFilterInternal(RequestContextFilter.java:100) - at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:119) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.springframework.web.filter.FormContentFilter.doFilterInternal(FormContentFilter.java:93) - at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:119) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:201) - at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:119) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:202) - at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) - at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:526) - at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:139) - at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:92) - at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:74) - at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343) - at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:408) - at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:66) - at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:861) - at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1579) - at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) - at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) - at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) - at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) - at java.lang.Thread.run(Thread.java:748) +2022-05-19 18:04:51.528 [main] ERROR o.s.b.diagnostics.LoggingFailureAnalysisReporter - -2022-05-05 13:35:46.902 [http-nio-8001-exec-6] ERROR c.a.s.exceptionhandler.GlobalExceptionHandler - GuliException(code=20001, msg=执行了自定义异常处理...) - at com.atguigu.eduservice.controller.EduTeacherController.findAllTeacher(EduTeacherController.java:54) - at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) - at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) - at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) - at java.lang.reflect.Method.invoke(Method.java:498) - at org.springframework.web.method.support.InvocableHandlerMethod.doInvoke(InvocableHandlerMethod.java:190) - at org.springframework.web.method.support.InvocableHandlerMethod.invokeForRequest(InvocableHandlerMethod.java:138) - at org.springframework.web.servlet.mvc.method.annotation.ServletInvocableHandlerMethod.invokeAndHandle(ServletInvocableHandlerMethod.java:106) - at org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.invokeHandlerMethod(RequestMappingHandlerAdapter.java:888) - at org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.handleInternal(RequestMappingHandlerAdapter.java:793) - at org.springframework.web.servlet.mvc.method.AbstractHandlerMethodAdapter.handle(AbstractHandlerMethodAdapter.java:87) - at org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:1040) - at org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:943) - at org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:1006) - at org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:898) - at javax.servlet.http.HttpServlet.service(HttpServlet.java:634) - at org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:883) - at javax.servlet.http.HttpServlet.service(HttpServlet.java:741) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:53) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.springframework.web.filter.RequestContextFilter.doFilterInternal(RequestContextFilter.java:100) - at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:119) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.springframework.web.filter.FormContentFilter.doFilterInternal(FormContentFilter.java:93) - at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:119) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:201) - at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:119) - at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) - at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) - at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:202) - at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) - at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:526) - at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:139) - at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:92) - at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:74) - at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343) - at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:408) - at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:66) - at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:861) - at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1579) - at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) - at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) - at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) - at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) - at java.lang.Thread.run(Thread.java:748) +*************************** +APPLICATION FAILED TO START +*************************** + +Description: + +Web server failed to start. Port 8001 was already in use. + +Action: + +Identify and stop the process that's listening on port 8001 or configure this application to listen on another port. diff --git a/guli_parent/service/service_edu/edu/log_info.log b/guli_parent/service/service_edu/edu/log_info.log index ba4edc5..470b701 100644 --- a/guli_parent/service/service_edu/edu/log_info.log +++ b/guli_parent/service/service_edu/edu/log_info.log @@ -1,172 +1,30 @@ -2022-05-06 09:52:35.727 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 15204 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-05-06 09:52:35.895 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-05-06 09:52:40.103 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-05-06 09:52:40.112 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-05-06 09:52:40.230 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 47ms. Found 0 repository interfaces. -2022-05-06 09:52:40.848 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-05-06 09:52:41.357 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-05-06 09:52:41.371 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-05-06 09:52:41.372 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-05-06 09:52:41.372 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-05-06 09:52:41.514 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-05-06 09:52:41.515 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 5147 ms -2022-05-06 09:52:41.986 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-05-06 09:52:42.365 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-05-06 09:52:43.652 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-05-06 09:52:43.906 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-05-06 09:52:44.986 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-05-06 09:52:45.020 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-05-06 09:52:45.065 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-05-06 09:52:45.511 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-05-06 09:52:45.533 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-05-06 09:52:45.536 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 13.212 seconds (JVM running for 15.645) -2022-05-06 09:53:56.118 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-05-06 09:53:56.118 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-05-06 09:53:56.148 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 30 ms -2022-05-06 10:57:01.474 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 12332 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-05-06 10:57:01.531 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-05-06 10:57:08.139 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-05-06 10:57:08.164 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-05-06 10:57:08.457 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 154ms. Found 0 repository interfaces. -2022-05-06 10:57:10.323 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-05-06 10:57:11.494 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-05-06 10:57:11.541 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-05-06 10:57:11.543 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-05-06 10:57:11.546 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-05-06 10:57:11.963 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-05-06 10:57:11.964 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 9710 ms -2022-05-06 10:57:12.883 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-05-06 10:57:13.569 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-05-06 10:57:16.737 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-05-06 10:57:17.184 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-05-06 10:57:19.150 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-05-06 10:57:19.209 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-05-06 10:57:19.303 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-05-06 10:57:20.172 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-05-06 10:57:20.220 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-05-06 10:57:20.238 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 23.142 seconds (JVM running for 26.723) -2022-05-06 11:08:35.770 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 13764 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-05-06 11:08:35.790 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-05-06 11:08:42.862 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-05-06 11:08:42.914 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-05-06 11:08:43.151 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 92ms. Found 0 repository interfaces. -2022-05-06 11:08:45.479 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-05-06 11:08:47.255 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-05-06 11:08:47.341 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-05-06 11:08:47.346 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-05-06 11:08:47.362 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-05-06 11:08:48.160 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-05-06 11:08:48.168 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 12056 ms -2022-05-06 11:08:49.566 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-05-06 11:08:50.610 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-05-06 11:08:54.309 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-05-06 11:08:54.972 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-05-06 11:08:57.158 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-05-06 11:08:57.254 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-05-06 11:08:57.443 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-05-06 11:08:58.668 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-05-06 11:08:58.749 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-05-06 11:08:58.770 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 26.387 seconds (JVM running for 29.882) -2022-05-06 11:09:05.928 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-05-06 11:09:05.931 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-05-06 11:09:05.965 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 31 ms -2022-05-06 11:11:22.419 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 14116 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-05-06 11:11:22.423 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-05-06 11:11:24.227 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-05-06 11:11:24.228 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-05-06 11:11:24.315 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 24ms. Found 0 repository interfaces. -2022-05-06 11:11:24.877 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-05-06 11:11:25.223 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-05-06 11:11:25.229 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-05-06 11:11:25.230 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-05-06 11:11:25.230 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-05-06 11:11:25.362 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-05-06 11:11:25.363 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 2671 ms -2022-05-06 11:11:25.631 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-05-06 11:11:25.828 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-05-06 11:11:27.023 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-05-06 11:11:27.268 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-05-06 11:11:27.873 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-05-06 11:11:27.882 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-05-06 11:11:27.926 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-05-06 11:11:28.213 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-05-06 11:11:28.233 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-05-06 11:11:28.235 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 7.039 seconds (JVM running for 8.971) -2022-05-06 11:11:32.024 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-05-06 11:11:32.026 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-05-06 11:11:32.033 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 5 ms -2022-05-06 11:14:30.219 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 11988 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-05-06 11:14:30.235 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-05-06 11:14:32.096 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-05-06 11:14:32.098 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-05-06 11:14:32.122 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 15ms. Found 0 repository interfaces. -2022-05-06 11:14:32.506 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-05-06 11:14:32.813 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-05-06 11:14:32.832 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-05-06 11:14:32.832 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-05-06 11:14:32.833 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-05-06 11:14:32.925 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-05-06 11:14:32.925 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 2594 ms -2022-05-06 11:14:33.236 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-05-06 11:14:33.448 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-05-06 11:14:34.377 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-05-06 11:14:34.519 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-05-06 11:14:35.107 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-05-06 11:14:35.116 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-05-06 11:14:35.171 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-05-06 11:14:35.483 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-05-06 11:14:35.510 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-05-06 11:14:35.512 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 6.722 seconds (JVM running for 7.686) -2022-05-06 11:14:40.991 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-05-06 11:14:40.992 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-05-06 11:14:41.010 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 18 ms -2022-05-06 11:17:35.715 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 8296 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-05-06 11:17:35.718 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-05-06 11:17:37.702 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-05-06 11:17:37.704 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-05-06 11:17:37.767 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 9ms. Found 0 repository interfaces. -2022-05-06 11:17:38.298 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-05-06 11:17:38.631 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-05-06 11:17:38.636 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-05-06 11:17:38.637 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-05-06 11:17:38.637 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-05-06 11:17:38.759 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-05-06 11:17:38.760 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 2849 ms -2022-05-06 11:17:39.045 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-05-06 11:17:39.303 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-05-06 11:17:40.176 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-05-06 11:17:40.307 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-05-06 11:17:40.969 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-05-06 11:17:40.978 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-05-06 11:17:41.041 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-05-06 11:17:41.421 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-05-06 11:17:41.442 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-05-06 11:17:41.444 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 6.944 seconds (JVM running for 8.254) -2022-05-06 11:17:48.500 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-05-06 11:17:48.501 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-05-06 11:17:48.509 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 8 ms -2022-05-06 11:18:43.986 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 15180 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) -2022-05-06 11:18:43.989 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev -2022-05-06 11:18:46.183 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! -2022-05-06 11:18:46.185 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. -2022-05-06 11:18:46.223 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 19ms. Found 0 repository interfaces. -2022-05-06 11:18:46.792 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) -2022-05-06 11:18:47.210 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) -2022-05-06 11:18:47.250 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] -2022-05-06 11:18:47.251 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] -2022-05-06 11:18:47.251 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] -2022-05-06 11:18:47.390 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext -2022-05-06 11:18:47.390 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3257 ms -2022-05-06 11:18:47.787 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... -2022-05-06 11:18:48.020 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. -2022-05-06 11:18:48.934 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] -2022-05-06 11:18:49.072 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' -2022-05-06 11:18:49.781 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed -2022-05-06 11:18:49.790 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) -2022-05-06 11:18:49.849 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references -2022-05-06 11:18:50.136 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] -2022-05-06 11:18:50.176 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat started on port(s): 8001 (http) with context path '' -2022-05-06 11:18:50.182 [main] INFO com.atguigu.eduservice.EduApplication - Started EduApplication in 7.938 seconds (JVM running for 10.394) -2022-05-06 11:18:54.742 [http-nio-8001-exec-1] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring DispatcherServlet 'dispatcherServlet' -2022-05-06 11:18:54.742 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Initializing Servlet 'dispatcherServlet' -2022-05-06 11:18:54.773 [http-nio-8001-exec-1] INFO org.springframework.web.servlet.DispatcherServlet - Completed initialization in 31 ms +2022-05-19 18:04:44.815 [main] INFO com.atguigu.eduservice.EduApplication - Starting EduApplication on SKY-20201118IYZ with PID 6140 (E:\self_example\guli_parent\service\service_edu\target\classes started by Administrator in E:\self_example\guli_parent) +2022-05-19 18:04:44.905 [main] INFO com.atguigu.eduservice.EduApplication - The following profiles are active: dev +2022-05-19 18:04:47.379 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Multiple Spring Data modules found, entering strict repository configuration mode! +2022-05-19 18:04:47.382 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Bootstrapping Spring Data repositories in DEFAULT mode. +2022-05-19 18:04:47.450 [main] INFO o.s.d.r.config.RepositoryConfigurationDelegate - Finished Spring Data repository scanning in 54ms. Found 0 repository interfaces. +2022-05-19 18:04:47.875 [main] INFO o.s.c.s.PostProcessorRegistrationDelegate$BeanPostProcessorChecker - Bean 'org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration' of type [org.springframework.transaction.annotation.ProxyTransactionManagementConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying) +2022-05-19 18:04:48.253 [main] INFO o.s.boot.web.embedded.tomcat.TomcatWebServer - Tomcat initialized with port(s): 8001 (http) +2022-05-19 18:04:48.262 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Initializing ProtocolHandler ["http-nio-8001"] +2022-05-19 18:04:48.262 [main] INFO org.apache.catalina.core.StandardService - Starting service [Tomcat] +2022-05-19 18:04:48.262 [main] INFO org.apache.catalina.core.StandardEngine - Starting Servlet engine: [Apache Tomcat/9.0.27] +2022-05-19 18:04:48.386 [main] INFO o.a.c.core.ContainerBase.[Tomcat].[localhost].[/] - Initializing Spring embedded WebApplicationContext +2022-05-19 18:04:48.387 [main] INFO org.springframework.web.context.ContextLoader - Root WebApplicationContext: initialization completed in 3323 ms +2022-05-19 18:04:48.732 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Starting... +2022-05-19 18:04:49.010 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Start completed. +2022-05-19 18:04:50.124 [main] INFO s.d.s.w.PropertySourcedRequestMappingHandlerMapping - Mapped URL path [/v2/api-docs] onto method [springfox.documentation.swagger2.web.Swagger2Controller#getDocumentation(String, HttpServletRequest)] +2022-05-19 18:04:50.271 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Initializing ExecutorService 'applicationTaskExecutor' +2022-05-19 18:04:50.941 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Context refreshed +2022-05-19 18:04:50.990 [main] INFO s.d.s.web.plugins.DocumentationPluginsBootstrapper - Found 1 custom documentation plugin(s) +2022-05-19 18:04:51.047 [main] INFO s.d.spring.web.scanners.ApiListingReferenceScanner - Scanning for api listing references +2022-05-19 18:04:51.467 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Starting ProtocolHandler ["http-nio-8001"] +2022-05-19 18:04:51.479 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Pausing ProtocolHandler ["http-nio-8001"] +2022-05-19 18:04:51.480 [main] INFO org.apache.catalina.core.StandardService - Stopping service [Tomcat] +2022-05-19 18:04:51.491 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Stopping ProtocolHandler ["http-nio-8001"] +2022-05-19 18:04:51.493 [main] INFO org.apache.coyote.http11.Http11NioProtocol - Destroying ProtocolHandler ["http-nio-8001"] +2022-05-19 18:04:51.515 [main] INFO o.s.b.a.l.ConditionEvaluationReportLoggingListener - + +Error starting ApplicationContext. To display the conditions report re-run your application with 'debug' enabled. +2022-05-19 18:04:51.602 [main] INFO o.s.scheduling.concurrent.ThreadPoolTaskExecutor - Shutting down ExecutorService 'applicationTaskExecutor' +2022-05-19 18:04:51.604 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Shutdown initiated... +2022-05-19 18:04:51.608 [main] INFO com.zaxxer.hikari.HikariDataSource - HikariPool-1 - Shutdown completed. diff --git a/phm_rotate/backend/phm_backend/collecting/pom.xml b/phm_rotate/backend/phm_backend/collecting/pom.xml new file mode 100644 index 0000000..8f0d7cf --- /dev/null +++ b/phm_rotate/backend/phm_backend/collecting/pom.xml @@ -0,0 +1,19 @@ + + + + phm_parent + com.cqu + 0.0.1-SNAPSHOT + + 4.0.0 + + pom + + mqtt-collect-flume + + collecting + + + \ No newline at end of file