🔧 Fixing broken AirDCPP search
This commit is contained in:
103
dependencies.docker-compose.yml
Normal file
103
dependencies.docker-compose.yml
Normal file
@@ -0,0 +1,103 @@
|
|||||||
|
services:
|
||||||
|
zoo1:
|
||||||
|
image: confluentinc/cp-zookeeper:7.3.2
|
||||||
|
hostname: zoo1
|
||||||
|
container_name: zoo1
|
||||||
|
ports:
|
||||||
|
- "2181:2181"
|
||||||
|
environment:
|
||||||
|
ZOOKEEPER_CLIENT_PORT: 2181
|
||||||
|
ZOOKEEPER_SERVER_ID: 1
|
||||||
|
ZOOKEEPER_SERVERS: zoo1:2888:3888
|
||||||
|
networks:
|
||||||
|
- kafka-net
|
||||||
|
|
||||||
|
kafka1:
|
||||||
|
image: confluentinc/cp-kafka:7.3.2
|
||||||
|
hostname: kafka1
|
||||||
|
container_name: kafka1
|
||||||
|
ports:
|
||||||
|
- "9092:9092"
|
||||||
|
- "29092:29092"
|
||||||
|
- "9999:9999"
|
||||||
|
environment:
|
||||||
|
KAFKA_ADVERTISED_LISTENERS: INTERNAL://kafka1:19092,EXTERNAL://${DOCKER_HOST_IP:-127.0.0.1}:9092,DOCKER://host.docker.internal:29092
|
||||||
|
KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: INTERNAL:PLAINTEXT,EXTERNAL:PLAINTEXT,DOCKER:PLAINTEXT
|
||||||
|
KAFKA_INTER_BROKER_LISTENER_NAME: INTERNAL
|
||||||
|
KAFKA_ZOOKEEPER_CONNECT: "zoo1:2181"
|
||||||
|
KAFKA_BROKER_ID: 1
|
||||||
|
KAFKA_LOG4J_LOGGERS: "kafka.controller=INFO,kafka.producer.async.DefaultEventHandler=INFO,state.change.logger=INFO"
|
||||||
|
KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 1
|
||||||
|
KAFKA_TRANSACTION_STATE_LOG_REPLICATION_FACTOR: 1
|
||||||
|
KAFKA_TRANSACTION_STATE_LOG_MIN_ISR: 1
|
||||||
|
KAFKA_JMX_PORT: 9999
|
||||||
|
KAFKA_JMX_HOSTNAME: ${DOCKER_HOST_IP:-127.0.0.1}
|
||||||
|
KAFKA_AUTHORIZER_CLASS_NAME: kafka.security.authorizer.AclAuthorizer
|
||||||
|
KAFKA_ALLOW_EVERYONE_IF_NO_ACL_FOUND: "true"
|
||||||
|
depends_on:
|
||||||
|
- zoo1
|
||||||
|
networks:
|
||||||
|
- kafka-net
|
||||||
|
|
||||||
|
kafka-ui:
|
||||||
|
container_name: kafka-ui
|
||||||
|
image: provectuslabs/kafka-ui:latest
|
||||||
|
ports:
|
||||||
|
- 8087:8080
|
||||||
|
environment:
|
||||||
|
DYNAMIC_CONFIG_ENABLED: true
|
||||||
|
volumes:
|
||||||
|
- /Users/rishi/work/config/kafka-ui/config.yml:/etc/kafkaui/dynamic_config.yaml
|
||||||
|
depends_on:
|
||||||
|
- kafka1
|
||||||
|
- zoo1
|
||||||
|
networks:
|
||||||
|
- kafka-net
|
||||||
|
|
||||||
|
db:
|
||||||
|
image: "mongo:latest"
|
||||||
|
container_name: database
|
||||||
|
networks:
|
||||||
|
- kafka-net
|
||||||
|
ports:
|
||||||
|
- "27017:27017"
|
||||||
|
volumes:
|
||||||
|
- "mongodb_data:/bitnami/mongodb"
|
||||||
|
|
||||||
|
redis:
|
||||||
|
image: "bitnami/redis:latest"
|
||||||
|
container_name: queue
|
||||||
|
environment:
|
||||||
|
ALLOW_EMPTY_PASSWORD: "yes"
|
||||||
|
networks:
|
||||||
|
- kafka-net
|
||||||
|
ports:
|
||||||
|
- "6379:6379"
|
||||||
|
|
||||||
|
elasticsearch:
|
||||||
|
image: docker.elastic.co/elasticsearch/elasticsearch:7.16.2
|
||||||
|
container_name: elasticsearch
|
||||||
|
environment:
|
||||||
|
- "discovery.type=single-node"
|
||||||
|
- "ES_JAVA_OPTS=-Xms512m -Xmx512m"
|
||||||
|
- "xpack.security.enabled=true"
|
||||||
|
- "xpack.security.authc.api_key.enabled=true"
|
||||||
|
- "ELASTIC_PASSWORD=password"
|
||||||
|
ulimits:
|
||||||
|
memlock:
|
||||||
|
soft: -1
|
||||||
|
hard: -1
|
||||||
|
ports:
|
||||||
|
- "9200:9200"
|
||||||
|
networks:
|
||||||
|
- kafka-net
|
||||||
|
|
||||||
|
networks:
|
||||||
|
kafka-net:
|
||||||
|
driver: bridge
|
||||||
|
|
||||||
|
volumes:
|
||||||
|
mongodb_data:
|
||||||
|
driver: local
|
||||||
|
elasticsearch:
|
||||||
|
driver: local
|
||||||
@@ -551,7 +551,8 @@ export default class ImportService extends Service {
|
|||||||
params: { id: "string" },
|
params: { id: "string" },
|
||||||
async handler(ctx: Context<{ id: string }>) {
|
async handler(ctx: Context<{ id: string }>) {
|
||||||
console.log(ctx.params.id);
|
console.log(ctx.params.id);
|
||||||
return await Comic.findById(ctx.params.id);
|
return await Comic.findById(new ObjectId(ctx.params.id));
|
||||||
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
getComicBooksByIds: {
|
getComicBooksByIds: {
|
||||||
|
|||||||
@@ -125,12 +125,9 @@ export default class SocketService extends Service {
|
|||||||
config: "object",
|
config: "object",
|
||||||
},
|
},
|
||||||
async handler(ctx) {
|
async handler(ctx) {
|
||||||
console.log("a, a kanha kanha...");
|
|
||||||
|
|
||||||
const { query, config, namespace } = ctx.params;
|
const { query, config, namespace } = ctx.params;
|
||||||
const namespacedInstance = this.io.of(namespace || "/");
|
const namespacedInstance = this.io.of(namespace || "/");
|
||||||
const ADCPPSocket = new AirDCPPSocket(config);
|
const ADCPPSocket = new AirDCPPSocket(config);
|
||||||
console.log("asdas", ADCPPSocket);
|
|
||||||
try {
|
try {
|
||||||
await ADCPPSocket.connect();
|
await ADCPPSocket.connect();
|
||||||
const instance = await ADCPPSocket.post(
|
const instance = await ADCPPSocket.post(
|
||||||
@@ -148,6 +145,7 @@ export default class SocketService extends Service {
|
|||||||
`search`,
|
`search`,
|
||||||
`search_result_added`,
|
`search_result_added`,
|
||||||
(groupedResult) => {
|
(groupedResult) => {
|
||||||
|
console.log(JSON.stringify(groupedResult, null, 4));
|
||||||
namespacedInstance.emit(
|
namespacedInstance.emit(
|
||||||
"searchResultAdded",
|
"searchResultAdded",
|
||||||
groupedResult
|
groupedResult
|
||||||
|
|||||||
Reference in New Issue
Block a user