2026-01-09 23:18:52 +08:00
|
|
|
import { Injectable } from '@nestjs/common';
|
|
|
|
|
import { InjectRepository } from '@nestjs/typeorm';
|
2026-01-12 12:28:37 +08:00
|
|
|
import { Repository, LessThan, MoreThanOrEqual } from 'typeorm';
|
2026-01-09 23:18:52 +08:00
|
|
|
import { BidItem } from '../entities/bid-item.entity';
|
2026-01-13 19:46:41 +08:00
|
|
|
import { CrawlInfoAdd } from '../../crawler/entities/crawl-info-add.entity';
|
2026-01-09 23:18:52 +08:00
|
|
|
|
|
|
|
|
@Injectable()
|
|
|
|
|
export class BidsService {
|
|
|
|
|
constructor(
|
|
|
|
|
@InjectRepository(BidItem)
|
|
|
|
|
private bidRepository: Repository<BidItem>,
|
2026-01-13 19:46:41 +08:00
|
|
|
@InjectRepository(CrawlInfoAdd)
|
|
|
|
|
private crawlInfoRepository: Repository<CrawlInfoAdd>,
|
2026-01-09 23:18:52 +08:00
|
|
|
) {}
|
|
|
|
|
|
|
|
|
|
async findAll(query?: any) {
|
|
|
|
|
const { page = 1, limit = 10, source, keyword } = query || {};
|
|
|
|
|
const qb = this.bidRepository.createQueryBuilder('bid');
|
|
|
|
|
|
|
|
|
|
if (source) {
|
|
|
|
|
qb.andWhere('bid.source = :source', { source });
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (keyword) {
|
|
|
|
|
qb.andWhere('bid.title LIKE :keyword', { keyword: `%${keyword}%` });
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
qb.orderBy('bid.publishDate', 'DESC')
|
|
|
|
|
.skip((page - 1) * limit)
|
|
|
|
|
.take(limit);
|
|
|
|
|
|
|
|
|
|
const [items, total] = await qb.getManyAndCount();
|
|
|
|
|
return { items, total };
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
async createOrUpdate(data: Partial<BidItem>) {
|
2026-01-13 00:39:43 +08:00
|
|
|
// Use title or a hash of title to check for duplicates
|
|
|
|
|
let item = await this.bidRepository.findOne({ where: { title: data.title } });
|
2026-01-09 23:18:52 +08:00
|
|
|
if (item) {
|
|
|
|
|
Object.assign(item, data);
|
|
|
|
|
return this.bidRepository.save(item);
|
|
|
|
|
}
|
|
|
|
|
return this.bidRepository.save(data);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
async cleanOldData() {
|
|
|
|
|
const thirtyDaysAgo = new Date();
|
|
|
|
|
thirtyDaysAgo.setDate(thirtyDaysAgo.getDate() - 30);
|
|
|
|
|
return this.bidRepository.delete({
|
|
|
|
|
createdAt: LessThan(thirtyDaysAgo),
|
|
|
|
|
});
|
|
|
|
|
}
|
2026-01-12 02:09:48 +08:00
|
|
|
|
|
|
|
|
async getSources() {
|
|
|
|
|
const result = await this.bidRepository
|
|
|
|
|
.createQueryBuilder('bid')
|
|
|
|
|
.select('DISTINCT bid.source')
|
|
|
|
|
.where('bid.source IS NOT NULL')
|
|
|
|
|
.orderBy('bid.source', 'ASC')
|
|
|
|
|
.getRawMany();
|
|
|
|
|
return result.map((item: any) => item.source);
|
|
|
|
|
}
|
2026-01-12 12:28:37 +08:00
|
|
|
|
|
|
|
|
async getRecentBids() {
|
|
|
|
|
const thirtyDaysAgo = new Date();
|
|
|
|
|
thirtyDaysAgo.setDate(thirtyDaysAgo.getDate() - 30);
|
|
|
|
|
thirtyDaysAgo.setHours(0, 0, 0, 0);
|
|
|
|
|
|
|
|
|
|
return this.bidRepository
|
|
|
|
|
.createQueryBuilder('bid')
|
|
|
|
|
.where('bid.publishDate >= :thirtyDaysAgo', { thirtyDaysAgo })
|
|
|
|
|
.orderBy('bid.publishDate', 'DESC')
|
|
|
|
|
.getMany();
|
|
|
|
|
}
|
2026-01-12 18:59:17 +08:00
|
|
|
|
2026-01-12 22:00:39 +08:00
|
|
|
async getBidsByDateRange(startDate?: string, endDate?: string, keywords?: string[]) {
|
2026-01-12 18:59:17 +08:00
|
|
|
const qb = this.bidRepository.createQueryBuilder('bid');
|
|
|
|
|
|
|
|
|
|
if (startDate) {
|
|
|
|
|
const start = new Date(startDate);
|
|
|
|
|
start.setHours(0, 0, 0, 0);
|
|
|
|
|
qb.andWhere('bid.publishDate >= :startDate', { startDate: start });
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (endDate) {
|
|
|
|
|
const end = new Date(endDate);
|
|
|
|
|
end.setHours(23, 59, 59, 999);
|
|
|
|
|
qb.andWhere('bid.publishDate <= :endDate', { endDate: end });
|
|
|
|
|
}
|
|
|
|
|
|
2026-01-12 22:00:39 +08:00
|
|
|
if (keywords && keywords.length > 0) {
|
|
|
|
|
const keywordConditions = keywords.map((keyword, index) => {
|
|
|
|
|
return `bid.title LIKE :keyword${index}`;
|
|
|
|
|
}).join(' OR ');
|
|
|
|
|
qb.andWhere(`(${keywordConditions})`, keywords.reduce((params, keyword, index) => {
|
|
|
|
|
params[`keyword${index}`] = `%${keyword}%`;
|
|
|
|
|
return params;
|
|
|
|
|
}, {}));
|
|
|
|
|
}
|
|
|
|
|
|
2026-01-12 18:59:17 +08:00
|
|
|
return qb.orderBy('bid.publishDate', 'DESC').getMany();
|
|
|
|
|
}
|
2026-01-12 22:00:39 +08:00
|
|
|
|
|
|
|
|
async getCrawlInfoAddStats() {
|
2026-01-13 19:46:41 +08:00
|
|
|
// 获取每个来源的最新一次爬虫记录(按 createdAt 降序)
|
2026-01-12 22:00:39 +08:00
|
|
|
const query = `
|
|
|
|
|
SELECT
|
|
|
|
|
source,
|
|
|
|
|
count,
|
|
|
|
|
latestPublishDate,
|
|
|
|
|
error,
|
|
|
|
|
createdAt as latestUpdate
|
|
|
|
|
FROM crawl_info_add
|
2026-01-13 19:46:41 +08:00
|
|
|
WHERE (source, createdAt) IN (
|
|
|
|
|
SELECT source, MAX(createdAt)
|
2026-01-12 22:00:39 +08:00
|
|
|
FROM crawl_info_add
|
|
|
|
|
GROUP BY source
|
|
|
|
|
)
|
|
|
|
|
ORDER BY source ASC
|
|
|
|
|
`;
|
|
|
|
|
|
2026-01-13 19:46:41 +08:00
|
|
|
const results = await this.crawlInfoRepository.query(query);
|
2026-01-12 22:00:39 +08:00
|
|
|
|
|
|
|
|
return results.map((item: any) => ({
|
|
|
|
|
source: item.source,
|
|
|
|
|
count: item.count,
|
|
|
|
|
latestUpdate: item.latestUpdate,
|
|
|
|
|
latestPublishDate: item.latestPublishDate,
|
|
|
|
|
error: item.error,
|
|
|
|
|
}));
|
|
|
|
|
}
|
2026-01-09 23:18:52 +08:00
|
|
|
}
|