Forecasting photovoltaic power in high-latitude regions via support vector machine optimized by meta-heuristics – Nature

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
Advertisement
Scientific Reports volume 16, Article number: 3438 (2026)
1053 Accesses
Metrics details
Machine-learning techniques are widely used across many disciplines, including electricity generation forecasting. In this study, the Support Vector Machine (SVM) based models, one of the machine learning techniques, were developed for daily PV power forecasting. To improve model performance, models were tuned with four metaheuristic optimizers, including the Artificial Bee Colony (ABC), Grey Wolf Optimizer (GWO), Genetic Algorithm (GA), and Particle Swarm Optimization (PSO). Daily PV power and temperature data from 2020 to 2023 were obtained for the Stavanger, Oslo, and Kristiansand regions which located in southern Norway. One of the innovative aspects of this study is the investigation of the performance of SVM (Support Vector Machine) combined with various optimization methods across four alternative input configurations. To examine the different feature combinations, four different input configurations were created through the Minimum-Redundancy Maximum-Relevancy (MRMR) method. The analysis results obtained with SVM were further enhanced using all optimization techniques. Among those, the SVM-PSO-M04 (r = 0.7707, NSE = 0.5748, KGE = 0.7092, PI = 0.2964 and RMSE = 0.6513) method produced the most effective results (improving the correlation coefficient (r) to 0.7707 (approximately a 19% increase over the untuned SVM)) among the tested hybrid configurations obtained in our experiments. Moreover, coupling temperature data alongside PV power as model input also tends to improve forecasting skill. Results of this study provide a case-study benchmark for researchers, institutions, and other stakeholders engaged in renewable energy planning and management in high-latitude regions.
Solar energy has emerged as a critical frontier of global decarbonisation strategies. Yet, under climate change, the power capacity of these systems remains fundamentally linked to meteorological conditions that are themselves undergoing a transformation1,2,3,4,5. Long term climate model ensembles consistently project regional scale shifts in both surface solar radiation and near-surface air temperature which jointly determine photovoltaic (PV) efficiency6,7,8,9,10. While solar energy is subject to geographic and temporal variations4,6, the relation between rising temperatures and solar energy production introduces a more complex technical challenge. The linear degradation of photovoltaic conversion efficiency as cell temperature rises is a fundamental issue particularly in urban or low-latitude regions which consequently can offset the benefits from increased insolation even with minor warming conditions9,11,12. A case study in Istanbul exemplifies this negative synergy. The study demonstrated that according to RegCM based projections through 2050, a small decrease in incoming radiation levels may pose a direct threat to future PV yields when coupled with simultaneous multi-degree warming13. For ensuring modern energy grids remain resilient in the face of a rapidly shifting climate accurate solar power forecasting is a fundamental necessity. 
Future climate conditions, intensified by climate change, are expected to increase the complexity of PV system from initial design through to operation, and maintenance. In Europe and parts of East Asia as the reduction in cloud cover is expected to offsets thermal performance losses, photovoltaic yields are expected to experience modest rise, even under aggressive warming scenarios4,5,7,9,10. Conversely regions like West Africa, North Africa, Australasia and parts of Central Asia are likely to face declines linked to solar dimming and rising cell temperatures, although most studies still find decreases remain within a 6% margin of current baselines4,6,8,10,12. While global case for solar remains as a no regrets investment, site specific planning must integrate a consistent climate signal into long-term energy planning to ensure sustained accuracy when projections are considered1,2,3,5,7.
Predictive analytics have become from a support tool to a fundamental need to maintain grid stability and manage net loads effectively as solar integration grows14,15. Beyond immediate operations, solar forecasting plays an important role in the economic profile of projects by optimizing battery storage and reducing wasted energy through curtailments16. It is suggested that integrating solar predictions with storage systems not only lowers operational expenses but also decreases overall grid dependency17. While various modeling techniques, apart from traditional regression, such as advanced deep learning, offer reliable short- and long-term outputs18, recent evidence suggests that the latter has demonstrated superior accuracy for irradiance forecasting19.Considering future climate where solar volatility is expected to intensify, the ability to quantify prediction uncertainty will be beneficial for microgrid resilience and the sophisticated management of modern power infrastructures20. Solar power prediction provides consistent grid operation, storage optimization, economic allocation, and integration of renewable sources. Furthermore, it helps in reducing the impacts of climate-induced variability. Ultimately, these forecasting tools serve as the backbone for consistent, climate-resilient power systems.
Recent research on solar-irradiance forecasting spans a wide spectrum of time-horizons and data sources, progressing from purely statistical baselines to highly integrated, image-enhanced deep-learning pipelines. Early work demonstrated the value of classic statistical approaches: Paulescu and Paulescu21 showed that an empirical two-state clear-sky model outperformed random-walk, moving-average and Autoregressive integrated moving average (ARIMA) baselines for four-samples-per-minute data from Timișoara, Romania, while Zambrano and Giraldo22 built multidimensional transfer models that dispense with costly on-site training measurements. Parallel semi-empirical efforts for hourly horizons combined extraterrestrial irradiance and clearness-index signals to surpass the Angström-Prescott formula at several Turkish sites23. Surveys and reviews have mapped the methodological landscape, covering statistical, cloud-image and NWP (Numerical Weather Prediction), routes24 and benchmarking time-series, image and hybrid families25.
Machine-learning studies have gradually pushed forecast granularity to the minute scale. Image-only pipelines link real-time sky-camera RGB profiles to 1–10 min global horizontal irradiance with competitive the mean absolute percentage error (MAPE) and the root mean square error (RMSE) scores26; all-sky imagers coupled with simultaneous irradiance readings achieved ramp-event detection indices of 43–62% at a Uruguayan test bed27. Satellite/NWP coupling also remains powerful at the 0–3 h range, trimming persistence errors by ≈ 10 W m⁻² across the U.S. SURFRAD network28. Where on-site imagery is unavailable, short-term physics-free predictors such as Artificial Neural Network (ANN)-SFP29 and daily-scale SVM/ANN/k-NN ensembles30 still yield R² values up to 0.94.
Deep learning has become the dominant trend for sub-hourly horizons. For example, CNN–LSTM hybrids (convolutional neural networks combined with long short-term memory networks) that fuse wavelet-packet-decomposed sequences with ground imagery have been shown to lower RMSE compared to back-propagation neural networks (BPNN), Support Vector Regression (SVR), and standalone LSTM models on three U.S. stations31, while multi-modal deep clustering aligns cloud-camera frames with the Numerical Weather Prediction (NWP) fields, reaching a 29.4 W m⁻² day-ahead RMSE in California32. Recurrent architectures remain strong: Deep Recurrent Neural Networks (DRNN) surpassed SVR and feed-forward networks in Canada33; LSTM networks delivered R² > 0.9 under complicated weather in Atlanta and Hawaii34; and bidirectional/attention LSTMs benefited from multi-site NASA POWER inputs across India35. Cutting-edge transformer variants now integrate variational-mode-decomposed components, eclipsing Convolutional Neural Network (CNN)-LSTM and vanilla transformer baselines over a 2015–2019 EMAP data set36. Comprehensive Indian reviews confirm that such CNN, LSTM and CNN-LSTM hybrids can lift accuracy by up to 37%37.
Collectively these studies highlight three converging insights for solar-radiation forecasting. First, hybridization, whether statistical-empirical, image-plus-NWP, or signal-decomposition-plus-deep-network, consistently boosts performance across climates. Second, model choice and horizon must reflect data availability: transfer learning and clear-sky filters remain valuable when imagery is scarce, whereas minute-ahead dispatch benefits most from sky cameras and CNN-LSTM fusion. Third, the field is shifting toward interpretable, multi-modal deep architectures that can generalize without local retraining, a direction reinforced by the superior accuracy of multi-modal deep clustering (MMDC)32 and Multiple Image Convolutional Long Short-Term Memory Fusion Network (MICNN-L)25. Remaining gaps include systematic cross-climate validation beyond the U.S., Europe and India, and unified uncertainty quantification to complement RMSE-based metrics.
In addition to new DL techniques and ML methods, nature-inspired metaheuristics have also emerged as powerful tools for solving complex nonlinear, multimodal optimization problems, particularly in the domains of machine learning model tuning, feature selection, and system design. Among these, the Artificial Bee Colony (ABC), Grey Wolf Optimizer (GWO), Genetic Algorithm (GA), and Whale Optimization Algorithm (WOA) are widely recognized for their balance of exploration and exploitation strategies. Those optimization algorithms are used both as a single optimization technique or a combination of various alternatives.
ABC algorithm, inspired by the foraging behavior of honeybees, has gained attention due to its simplicity and strong global search capabilities. As a population intelligent algorithm, ABC applied in many studies. Oruc et al.38, Zhang et al.39, Gujarathi et al.40 incorporated ABC directly employed with a ML method or hybridized with another optimization algorithm in studies ranging drought forecast to engine optimization and handling with high dimensional datasets. GWO mimics the social leadership and cooperative hunting strategies of grey wolves. It is valued for its algorithmic simplicity and convergence stability. GWO has been successfully applied in time series forecasting and parameter tuning tasks and demonstrated its effectiveness in optimizing41,42,43. GA is a classical evolutionary algorithm based on principles of natural selection, crossover, and mutation44. It has broad application across optimization tasks and is often used as a baseline for performance comparisons with newer methods. Genetic Algorithms remain to be used in hybrid models to boost convergence rates across problem spaces containing either discrete or continuous parameters in diverse application44,45,46,47. The Whale Optimization Algorithm (WOA) inspired from feeding strategy of whale population48. WOA gained recognition for its simplicity and ability to balance intensification and diversification. Tang et al.49 introduced a combined WOA-ABC algorithm to overcome local optima problem and improve solution accuracy across theoretical and practical engineering problems. Although less frequently cited in academic databases, the Coyote Optimization Algorithm (COA) or COATI represents another nature-inspired method that modeled on coyote pack dynamics and social learning50. Despite showing potential across broad domains ranging from energy planning to photovoltaic parameter extraction51,52, additional comparative studies are needed to validate its performance relative to better-established algorithms like ABC, GWO, and WOA. Similar to other algorithms, since its initial development, COA has evolved into several improved forms, including Chaotic COA (CCOA)53 and Multiobjective COA (MOCOA)54 which have been implemented in a broad range of problem domains.
As Norway targets toward broadening its energy mix and reduce its dependence on fossil fuels sources, solar power becomes the increasing component of the energy strategy. Norway now recognizes the importance of solar energy as an important complementary source of renewable electricity generation while it has been historically known for its hydropower capabilities55,56,57. The possibility of using solar energy across Norway’s diverse landscapes, such as urban areas, agricultural regions, and industrial sites, has become much more realistic due to the significant drop in solar PV system prices and progress in solar technology in recent years56,58. As solar energy will not be sufficient to supply Norway especially during the winter months, hybrid systems offer a promising solution for integrating solar energy into Norway’s energy landscape to meet the demands for renewable energy59. While Norway is not typically known for its solar energy production, it possesses a number of hidden advantages for PV performance. The favorable effect of low ambient temperatures on solar panel efficiency counterbalance lower irradiance levels by reducing heat-related efficiency losses since PV systems, especially those based on crystalline silicon, are thermally sensitive—their efficiency decreases by ~ 0.4–0.5% per °C increase in module temperature59,60,61,62,63. Moreover, new applications like icephobic nanocoatings64, considering climate variations and orientations affect65 increasing performance of the systems. Rees et al.,66 also built a simple, transparent workflow that turns freely available high-resolution LiDAR into city-wide rooftop-solar potential estimations and concluded that Rooftop solar PV in Tromsø could realistically supply ≈ 20–30% of the city’s annual electricity demand, about 200 GWh yr⁻¹, with residential roofs alone contributing roughly 40% of that total. In addition, new technologies, including tilting systems, bifacial panels, and heat recovery-integrated PV (PVT), present opportunities to improve year-round utilization67,68,69,70.
Given the evolving scientific landscape and Norway’s focus on solar energy, there is a need to integrate advanced solar power forecasting methodologies into regional energy planning and grid management. International studies have exhibited the superiority of hybrid deep learning models, especially those combining sky imagery, NWP outputs, and statistical decomposition. Though, local adoption in high-latitude contexts like Norway remains limited. The seasonal extremes in insolation, coupled with complex topography and urban form, demand forecasting frameworks that are both data-adaptive and climatically robust. Moreover, with evidence that rooftop solar in cities like Tromsø could meet up to 30% of local electricity demand, the importance of precise, site-specific irradiance forecasting becomes even more pronounced.
Many forecasting studies in environmental sciences and renewable-energy applications focus on a small set of established learning algorithms (e.g., ANN, SVM/SVR, and Random Forest) for which performance and generalization properties have been widely discussed71. In addition, several studies report that hybridization, through data preprocessing and/or metaheuristic hyperparameter tuning, can further improve SVM/SVR performance in related forecasting tasks72,73,74 Accordingly, the present study uses SVM as a controlled base learner to systematically quantify the incremental value of (i) feature selection and lag-structure design and (ii) different metaheuristic optimizers. The aim is not to claim universal superiority over all alternative machine-learning/deep-learning methods, but to provide a transparent, like-for-like comparison within a unified modeling framework.
In line with this objective, we develop and evaluate SVM-based hybrid forecasting models for daily PV power in southern Norway (Kristiansand, Stavanger, and Oslo) using PVGIS-ERA5 derived PV output and near-surface air temperature data covering 2020–2023. Four metaheuristic optimizers (ABC, GWO, GA, and PSO) are compared, and four alternative input configurations are defined via the Minimum Redundancy Maximum Relevance (MRMR) feature selection method. Model performance is assessed using multiple goodness-of-fit criteria (r, RMSE, NSE, KGE, and PI) together with visual diagnostics.
The main contributions of this work are as follows:
A daily PV power forecasting case study for three southern Norway sites using consistent PVGIS-ERA5 input data.
Development of a controlled comparison of four metaheuristic optimizers (ABC, GWO, GA, PSO) for tuning an SVM forecasting model.
Construction of four lagged input structures via MRMR and evaluation of four lagged input structures combining PV history and temperature signals.
Multi-metric evaluation (r, RMSE, NSE, KGE, PI) complemented by visual diagnostics to interpret performance differences and model limitations.
This work presents an analysis of PV generation prediction based on the data derived from three distinct locations of Norway. The analysis used SVM with a range of optimization techniques implemented including PSO, GWO, GA and ABC. The methodological framework in this study is summarized graphically in Fig. 1.
Methodological framework.
This study focuses on three coastal locations across southern Norway to evaluate solar radiation and PV power potential. The selected sites vary in latitude, elevation, and orientation. This diversification provides examination of how geographic and topographic parameters affect solar energy production. Geographically, the study area extends between 58.15° to ~ 59.90°N latitude and 5.73° to ~ 10.74°E longitude which covers a diverse range of coastal terrains.
Table 1 presents the key site characteristics including geographic location, elevation above sea level, slope, azimuth, and PV system specifications considered. The PVGIS-ERA5 radiation database was used to ensure climatic input consistency across all locations (PVGIS 5.3)75. The optimal slope angles fall within 46° to 49 range and azimuth values span from − 5° to 0°. These values indicate minor deviations from due south alignment, which is considered the optimal for fixed PV installations at Northern Hemisphere latitudes.
A standard monocrystalline silicon PV system with a nominal capacity of 1.0 kWp were assumed for each site. Uniform system losses of 14% are applied to account for thermal, wiring, and inverter-related inefficiencies. The selection of these representative sites and configuration aim to make the results more comparable and transferable to broader PV deployment scenarios across similar latitudes and climates.
In this study, hourly PV output and temperature records were first obtained and then aggregated to a daily scale for forecasting. Hourly PV output contains structurally zero values during nighttime; if retained, these long zero sequences can dominate error metrics and encourage trivial predictions dominated by nocturnal conditions rather than the physically informative daylight signal. To focus on daytime generation, nighttime hours were excluded and the daily PV target was computed as the total PV power generated during a representative 7-hour daylight window (selected to consistently capture peak sun hours for the region and to represent daytime generation consistently across the study period). Daily mean air temperature was used as the meteorological predictor. This aggregation inevitably reduces intra-day variability; therefore, sub-daily (hourly) forecasting and uncertainty quantification are recommended directions for future work. Table 1 represents statistics of data used in this study (Fig. 2).
Study locations in southern Norway.. The map was generated using ArcMap (version 10.8) from ArcGIS Desktop (https://desktop.arcgis.com).
Using different model structures in analyses may either increase or decrease model performance. Numerous studies and applications related to this topic can be found in the literature. Researchers often prefer feature selection methods grounded in statistical or stochastic process theory, and MRMR is one such technique. The Minimum Redundancy Maximum Relevance (MRMR) method is an effective feature selection technique designed to identify the optimal subset of input features for predicting an output. Its primary goal is to select features that are highly relevant to the output while maintaining minimum redundancy among themselves76. By prioritizing crucial, uncorrelated features, MRMR can enhance machine learning model accuracy and significantly reduce the risk of overfitting based on a greedy algorithm and a relevance-redundancy measure, making it particularly effective for high-dimensional datasets. In this study, the MRMR method was used to define the input structure for our models, based on its preference in the literature for yielding significant results77. For more details on the method, see Ding and Peng76. Through the MRMR, lagged input structures from PV output and air temperature were constructed. The significance levels and lagged values of temperature and PV data at different time intervals. Figure 3 shows the feature importance ranking from the MRMR analysis, and the resulting input combinations are summarized in Table 2.
Feature importance results from the MRMR analysis. (Pt = PV target value on day t; Pt5 = PV value lagged by 5 days; Tt4 = temperature value lagged by 4 days; etc.).
Support Vector Machines (SVMs) and their regression variant SVR rely on kernel functions to map nonlinear relationships in complex datasets. The support vector machine (SVM) is a classifier that belongs to the kernel approaches in machine learning. This learning system is employed to classify and predict the data fitness function, aiming to minimize mistakes in data categorization or the fitness function itself. To advance these methods not only the optimization kernel architectures or refining hyperparameters but also tailoring implementations for particular application contexts were used78,79,80,81. Wang et al.82 demonstrated that the Gaussian kernel’s spread parameter (γ) must fall within a specific range to ensure the model achieve optimal generalization. SVM is extensively utilized for both regression and classification problems. Owing to its adaptability and efficacy it is positioned as a premier method in machine learning. For practical applications, Kusuma and Kudus83 describe SVR as a regression method designed to control overfitting and demonstrate its application on mortgage survival data using a linear kernel. Theoretical foundations connecting SVMs to probabilistic frameworks emerged through Wang et al.‘s84 work linking Gaussian kernel density estimation (GKDE) to kernel-based learning. Their analysis revealed that Gaussian-kernel SVMs operate as probabilistic classifiers, thereby providing Bayesian justification for the algorithm’s empirical success.
.
The mathematical formulation of SVM is presented in Eq. (1), defining the relationship between input and output variables as follows:
where (:varphi:left(xright)) denotes a high dimensional feature space, w represents weight vectore, and b referred as the bias term. For implementation details and an extended reference trail used in closely related hybrid SVM studies the readers can refer to previous studies of Oruc et al.38 and Oruc et al.85 and for foundational SVR/SVM theory and formulation, cite standard SVR/SVM sources.
In this study, SVM algorithm was used to predict PV. Optimization methods, which are ABC, GWO, GA, and PSO, were used to enhance model performances. Initially, algorithm learning was performed with 70%, this ratio is frequently preferred in the literature, and it has been reported by many researchers that it yields effective results, of the dataset obtained from the region. Besides, time series data was not shuffled during the analysis because the sequence of data is critically important in time series analysis. The data was separated into training and testing sets without altering this sequence. The split ratios were selected based on proportions commonly found in the literature73,77,85. No random shuffling was applied, in order to simulate real forecasting conditions. Furthermore, preliminary steps or precautions against overfitting were taken using performance metrics like PI. Then, the optimization techniques mentioned above were used to further improve the performance of these algorithms. The results obtained from these models were evaluated according to performance evaluation criteria, RMSE, r, PI, KGE and NSE. All models were trained on the training set, and their performance was evaluated on the test set as described below.
Karaboga86 modeled the Artificial Bee Colony (ABC) algorithm based on honeybee foraging behavior. The algorithm divides the colony into three functional groups which exhibit distinct search behaviors. Employed bees exploit known food sources and communicate their quality through waggle dances at the hive. Onlooker bees observe these dances, evaluate potential sources based on profitability, and focus search efforts on those promising locations. Scout bees are responsible for exploring search space randomly and identifying new sources when existing sites are depleted87,88,89.
The algorithm maintains a balance between exploration and exploitation through dynamic role transitions. When a particular food source fail to show improvement over successive iterations, the employed bee assigned that location abandons it and transitions into a scout bee. This conversion starts random exploration for new opportunities90. This abandonment mechanism prevents the algorithm from being trapped in suboptimal solutions. The scouts introduce diversity and novelty into the search process, employed and onlooker bees refine and extract benefit from known solutions. Mathematical details and implementation procedures provided in Karaboga86, Li et al.88, and Vitorino et al.89.
Mirjalili et al.91 introduced the Grey Wolf Optimizer (GWO) by modeling the hunting dynamics observed in grey wolf packs. The algorithm transforms this behavior into a computational optimization framework. Grey wolves hunt through a hierarchical system of coordinated roles that includes tracking prey, encircling, and attacking. This is a strategy that translates effectively into computational search patterns92,93. The straightforward linear structure of the algorithm simplifies implementation while maintaining performance and has produced successful results in many fields, as mentioned by numerous researchers94.
GWO establishes candidate solutions based on the observed wolf pack hierarchy. The alpha (α) represents the current best solution and positions itself as the leader of the search process. Beta (β) and delta (δ) wolves correspond to the second and third-best solutions that guide exploration of potential regions. All remaining solutions function as omega (ω) wolves that represents the pack’s lowest tier that explores the broader search space91,95. This hierarchical structure drives the optimization mechanism forward. Alphas direct the hunt, betas and deltas refine the search direction and omegas ensure diversity in exploration.
Mathematically, the three hunting phases translate into iterative position adjustments. These are namely, tracking, encircling, and attacking phases which are guided by the alpha, beta, and delta solutions. Each iteration recalculates and refines wolf positions based on their distance from these leaders. The process progressively tightening the search around optimal regions until the algorithm reaches convergence, that is two criteria are determined in the GWO process: (1) catching the prey (reaching the best solution) and (2) reaching the maximum number of iterations96. Details governing the process can be found in Mirjalili et al.91.
Holland introduced the Genetic Algorithm (GA) in 1960 and refined it over the following two decades. The algorithm translates Darwin’s evolutionary principles into computational optimization methods. GA maintains a population of potential solutions that compete, recombine, and randomly mutate across iterative generations. This process directly parallels biological evolution where advantageous characteristics persist through populations while disadvantageous ones disappear44 and97. Industrial applications have validated the effectiveness of the algorithm across diverse application domains, ranging from parameter optimization to complex scheduling problems46,47,98,99.
The fundamental strength of the algorithm lies in its gradient-free search mechanism. This feature enables it to handle both continuous and discrete optimization challenges without requiring gradient calculations. Such a requirement constrains many traditional optimization methods44 and97. Bras et al.100 documented GA’s flexibility through applications extending from linguistic analysis to fuzzy network tuning. Their work demonstrates how the same evolutionary mechanisms, namely crossover, mutation, and selection, can be adapted to different problem domains100,101.
The algorithm proceeds through three core stages. First it initializes a random population, second applies genetic operators through probabilistic rather than deterministic process, and third evaluates solution quality until convergence criteria trigger termination101,102. When solutions fail to meet established quality thresholds, the algorithm iterates through another generational cycle. This iteration systematically refines and enhances the population through selection pressure and genetic variation101,103.
These steps are:
Crossover (stochastic): part of two solutions “is swapped” to produce new ones.
Mutation (stochastic): part of a new solution “is flipped” to generate a new one and prevent it from converging into local optima.
Selection: the new solutions are evaluated according to the objective function, and the best candidates are selected.
In certain instances, such as a high mutation rate that could lead to the loss of good solutions, the elitism operator is employed to guarantee that the optimal solutions are transferred to the next generation without modification, ensuring that the best candidates are maintained within the solution set100,103.
Kennedy and Eberhart introduced particle swarm optimization (PSO) in 1995. This algorithm conceptually inspires from coordinated and collective animal behaviors observed in natural system such as schools of fish navigating currents and bird flocks foraging collectively. Unlike genetic algorithms and other comparable evolutionary methods, PSO achieves more rapid convergence while requiring fewer computational resources. This advantage becomes particularly evident when addressing nonlinear optimization problems and impact modeling success104,105. The algorithm treats each potential solution as a particle moving through complex search spaces and is to use information of the current position X and velocity V of particles106.
Initially, particles are positioned randomly across the solution space. As the algorithm progresses through successive iterations, each particle adjusts its trajectory based on two guiding influences. These are the particle’s best position discovered so far and the globally best position identified by any particle of the entire swarm. This dual-component memory system drives particles toward potential solution regions while simultaneously maintaining exploration capability107,108,109,110,111. The search continues until particles converge on a shared optimal location. This convergence signals that the algorithm has identified the best available solution within the search space112,113,114.
The correlation coefficient (r), root mean square error (RMSE), Nash-Sutcliffe efficiency (NSE), Kling-Gupta efficiency (KGE), and Performance Index (PI), were employed key statistical indicators to assess model performance as specified in Equations (2) through (6). (2), (3), (4), (5), and (6) NSE and KGE are goodness-of-fit measures that attain 1 for perfect prediction. For NSE, a value of 0 indicates performance equivalent to the mean-prediction baseline, whereas KGE uses a benchmark threshold of approximately − 0.41 for comparable interpretation115.
In this study, total daily PV power generation data and daily average temperature values obtained from Kristiansand, Stavanger, and Oslo, which are in the southern region of Norway, were used to investigate the performance of forecasting models that were developed through combination of machine learning and optimization techniques. A total of 70% of the data was employed for training, while the remaining 30% was reserved for testing. Support Vector Machines (SVM) were selected as the machine learning algorithm, whereas optimization techniques included the Artificial Bee Colony (ABC), Grey Wolf Optimizer (GWO), Genetic Algorithm (GA), and Particle Swarm Optimization (PSO) due to one of the primary objectives of the study which is to enhance model performance by hybridizing. Furthermore, in order to examine the effect of model input variables on prediction performance, the MRMR (Minimum Redundancy Maximum Relevance) method was applied to construct four different input structures. The results of all analyses are presented in Table 3. Furthermore, it should be noted that these results and rankings (e.g., PSO-M04 being best) are specific to the data and sites analyzed; generalizing beyond this case study should be done with caution.
According to the results presented in Table 3, the performance metrics of the four input structures analyzed with SVM in the Kristiansand region differ from one another. Among these, the lowest performance was observed in M01, while the strongest baseline (untuned) SVM results were obtained with M04 (r = 0.6469, NSE = 0.3587, KGE = 0.3141, PI = 0.2937, and RMSE = 0.8004). The analysis of model structures shows that M01 was built using one input variable (Pt-5), while M04 was built using four input variables (Pt-5, Pt-3, Tt-4, and Tt-1). While both models used Pt as the target variable, integrating power data with temperature improves daily forecasting precision based on the results. Evidence from the analysis also reveals the success of MRMR-based input selection process in identifying superior input combinations. MRMR-based input selection process when coupled with hybridization of SVM with metaheuristic optimizers, also a consistent improvement was obtained in performance metrics (Table 3). Therefore, results also validate the value of hybridizing machine learning with optimization techniques compared to standalone configurations. However, the differences were marginal, which makes it challenging to identify a single superior optimizer for this region. Despite this, a closer look at the performance metrics revealed that SVM-PSO-M04 delivered the best test results for the Kristiansand region (r = 0.7707, NSE = 0.5748, KGE = 0.7092, PI = 0.2964, and RMSE = 0.6513). Among the tested models it can be concluded that metaheuristic hyperparameter tuning can enhance SVM performance for forecasting in this study.
Compared to the analysis without any optimization technique (SVM-M04), the performance metrics were improved. Additionally, when the results of the other optimization methods were evaluated, the M04 model consistently demonstrated success across all cases. This finding further validates the effectiveness of the MRMR method in determining the model input structure for this region; while these results should be interpreted as site- and dataset-specific.
Stavanger is another station within the study area where analyses were conducted. At this station, the same methods as in the previous case were applied, and the results are presented in Table 3. In the analyses performed solely with SVM, the M04 model achieved the highest scores among the four input structures (r = 0.6091, NSE = 0.3018, KGE = 0.2543, PI = 0.3548, and RMSE = 0.8352). Similar to the previous station, the input structure of M04 also demonstrated effective performance in this part of the analysis. For this station as well, optimization techniques were utilized to improve model performance. In the analyses involving ABC, GWO, GA, and PSO, performance metrics generally improved, with the results being very close to each other. Therefore, it was difficult to identify a single best-performing algorithm for the Stavanger station also. A primary takeaway of the analyses can be the overall performance among the optimization algorithms which produced comparable results. Another key observation is while the highest score metrics were obtained from M03 with the SVM-ABC and SVM-PSO techniques, all other methods achieved their highest scores using M04. These results suggest that the effectiveness of model input structures may vary upon regional environmental characteristics and they are sensitive to those characteristics. An additional finding is that using an optimal number of input variables with expanded data diversity which means balancing input complexity with representativeness, can achieve higher predictive precision.
Another station from which data were collected within the study area is Oslo. All of these stations are located in the southern region of Norway, where sunshine duration is considerably higher compared to other regions located in the country. The analysis results obtained from this station are presented in Table 3. As in the other stations, analyses were first performed using only SVM, and optimization techniques were subsequently applied to improve the results. Overall, the analyses incorporating optimization techniques improved the performance metrics of all models. In the analyses performed solely with SVM, the highest scores were obtained with M04 (r = 0.5994, NSE = 0.3043, KGE = 0.2518, PI = 0.3344, and RMSE = 0.8336). Among the optimization-based analyses, SVM-PSO-M04 achieved the highest test scores (r = 0.6861, NSE = 0.4470, KGE = 0.6252, PI = 0.3686, and RMSE = 0.7428), although differences among the top optimizers were modest. Unlike the Stavanger station, the best-performing input structure at this station was obtained with M04. Therefore, it can be concluded that while determining model input structures, not only PV power data but also temperature information can be beneficial for daily PV forecasting.
In general, across all regions, in this dataset, the strongest results were most often obtained with M04, while the input structure M01 exhibited the lowest performance metrics. All optimization techniques improved the performance of the models relative to the untuned SVM baseline. However, because performance differences among optimizers were relatively small, PSO should be interpreted as yielding the highest scores in this study rather than as universally superior.
In this study, in addition to statistical calculations, visual comparison methods were also employed. For the three different stations within the study area, violin plots were generated and are presented in Fig. 4. These plots were constructed to enable the comparison of the best-performing models within each category. According to the results, the only model that did not visually align with the observed values for Kristiansand was SVM-M04. Almost all of the other plots displayed highly similar distributions. Therefore, the most critical aspects to consider in this context are measures such as the mean, extremes, and median.
Violin diagram for all models in different stations on study area where SVM-M04; analysis of baseline SVM with M04 and, SVM-ABC-M04; analysis of SVM coupled ABC (hybridized SVM) with M04 etc. (a) Kristiansand (b) Stavanger (c) Oslo.
Although it is rather difficult to distinguish between the models based on visual observations alone, an evaluation of kernel densities, the 3rd quartile, and median values revealed that the model most similar to the observed data was SVM-PSO-M04. In Stavanger, the best-performing model was identified as SVM-PSO-M03, while in Oslo, the most successful model was again SVM-PSO-M04. In determining these successful models, multiple statistical parameters from the violin plots were taken into consideration.
In Fig. 5, a box-normal plot is presented to enable a better comparison of the best-performing models. Upon detailed examination of this plot, it appears quite difficult to distinguish the differences between the observed values and the prediction models. This difficulty mainly arises from the fact that all optimization methods produced results that were very close to each other. While the prediction models in the Kristiansand and Stavanger regions yielded highly similar outcomes, a few of the prediction models in the Oslo region could be distinguished more easily. In this region, the SVM-ABC-M03 and SVM-GWO-M04 models visually resembled the observed values more closely than the others; however, in terms of statistical results, these two methods lagged behind. Thus, the visual outcomes and the statistical results did not fully align. In conclusion, when the performance metrics obtained from prediction models are very close to one another, the box-normal method should not be preferred for visual comparison.
Box-normal diagram for all models in different stations on study area where SVM-M04; analysis of SVM with M04, SVM-ABC-M04; analysis of SVM and ABC with M04 etc. (a) Kristiansand (b) Stavanger (c) Oslo.
The ridge plot, another visual comparison method, is also included in this study. Figure 6 shows the ridge plot for all models that were successful in their respective class. Upon examining this plot, a model that visually overlaps with the observation values in the Kristiansand and Oslo regions could not be identified. However, in the Stavanger region, the models were able to predict the peak that occurred in the initial values of the observation data. But this was also not effective in determining the most successful model. Consequently, this visual comparison method cannot be considered successful either. In conclusion, the ridge plot was also unsuccessful in determining the best model, that is, in comparing prediction models whose performance metrics were very close to each other.
Ridge diagram for all models in different stations on study area where SVM-M04; analysis of SVM with M04, SVM-ABC-M04; analysis of SVM and ABC with M04 etc. (a) Kristiansand (b) Stavanger (c) Oslo.
Apart from the Violin plot, the visual comparison methods mentioned above did not yield exceptional results in distinguishing the most successful models. Therefore, to both increase the comparison comprehensiveness of the study and to enable a more qualified distinction of the results, Bland-Altman and Box plots were generated for all regions. All plots, specific to different regions, are shown in Figs. 7 and 8, and 9.
Bland-Altman diagram for all models in different stations on study area where SVM-M04; analysis of SVM with M04, SVM-ABC-M04; analysis of SVM and ABC with M04 etc. in Kristiansand.
In Fig. 7, the differences between the prediction models based on the observed values at 95% limits of agreement level are shown. One of the most important results here is that the differences for the SVM-PSO-M04 model are notably clustered around the zero-line indicating little bias for the SVM-PSO-M04 model. Although the predicted values for the other models also concentrate around zero, the densest region was determined to be SVM-PSO-M04. Therefore, this is an indication that it is the most successful model in that region.
Bland-Altman diagram for all models in different stations on study area where SVM-M04; analysis of SVM with M04, SVM-ABC-M04; analysis of SVM and ABC with M04 etc. in Stavanger.
Figure 8, on the other hand, shows the differences between the prediction models for Stavanger at 95% limits of agreement level, based on the observed values. The key point is that the predicted values for all models failed to cluster around the zero line, meaning their predictions have larger deviations from observations. Therefore, this indicates that no conclusive results were found for Stavanger. However, when looking at the SVM-PSO-M03 model, the predicted values are clustered between the 95% limits of agreement, which suggests that results close to the observed values were obtained. The predicted values are more concentrated between these two lines compared to the other models.
Bland-Altman diagram for all models in different stations on study area where SVM-M04; analysis of SVM with M04, SVM-ABC-M04; analysis of SVM and ABC with M04 etc. in Oslo.
Figure 9 displays the Bland-Altman plots for the prediction models based on the observed values for the Oslo station. Upon examining the figure, it is observed that the differences in the predicted values for the SVM-GA-M03 and SVM-PSO-M04 models are concentrated around zero. This indicates that these models yield more effective results compared to the others. The results here are consistent with the statistical findings.
To ensure the reliability of the results obtained in this study, ANOVA (Analysis of Variance) and Kruskal-Wallis statistical tests were applied to the forecasting models developed for the Stavanger, Kristiansand, and Oslo regions. The findings were examined at a 95% significance level in Table 4. For all stations, p-values were greater than 0.05, indicating that we fail to reject the null hypothesis of no statistically significant differences among the compared groups at 95% confidence. This suggests that the performance differences among the best models are modest and should be interpreted alongside the multi-metric evaluation and visual diagnostics. Also supporting the claim that while PSO-M04 had the highest metrics in general, its edge was not statistically significant.
The use of machine learning and optimization techniques in a hybridized manner is among the methods preferred by researchers across many disciplines in literature. Model performance metrics are typically improved using optimization techniques. In this study, consistent but moderate improvements in model performance values were also achieved through optimization techniques. The results obtained in this study, both in terms of the methods used and the optimization techniques applied, are consistent with those of several studies in literature. Some of these include: AlMohimeed et al.116 developed prediction models for the forward-looking estimation of cancer cells by utilizing image processing methods. The SVM-PSO model achieved one of the successful results. Just as this hybrid method was effective in the forward-looking prediction of cancer cells, it has also proven effective in the early diagnosis of hypertension problems117. In addition to their use in the health sector, hybrid methods are also utilized in hydrological studies. Oruc et al.38 investigated the performance of forward-looking drought prediction models in the Norwegian region by utilizing The Adaptive Neuro-Fuzzy Inference System (ANFIS) and SVM. They improved the model performances by hybridizing the machine learning algorithms with GWO, ABC, PSO, and GA. In their results, they emphasized that effective outcomes were obtained in analyses using SVM-PSO. In addition to these, they created 12 different model input structures using cross-correlation. In their conclusions, they stated that the data type and the number of delayed data points in the model input structure should be kept at an optimum level. Results consistent with these findings were also obtained in this study. To broaden the scope of the study, the model input structure in this work was created using the MRMR method instead of cross-correlation. Another study that obtained effective results with SVM-PSO is that of Samantaray et al.118. In their study, they chose Back Propagation Neural Network (BPNN) and SVM, along with PSO from the optimization techniques. Samantaray et al., who aimed to model the forward-looking prediction of floods in the Barak valley by creating 5 different input structures, obtained the most effective results from analyses performed with SVM-PSO. Unlike this study, they did not use any delayed data in their model input structures. Furthermore, they enriched their model input structure with meteorological data. In the results of their work, they also mentioned that effective results were achieved through this enrichment.
In many studies in the literature where hybrid machine learning and optimization techniques are preferred, usually only two or three optimization techniques are used. In this study, however, SVM, which is considered a well-established machine learning algorithm, was chosen and hybridized with four different optimization techniques (ABC, GWO, GA, and PSO). While these optimizers have been widely applied across different domains, side-by-side comparisons within a unified high-latitude PV forecasting setup remain relatively limited. Although the combined use of these methods strengthens the comparative aspect of our study, including additional machine learning models (e.g., a neural network or ensemble method) could further broaden the scope of comparison. However, such analyses involving optimization techniques are highly time-consuming and computationally expensive. This would move beyond the scope of this study and potentially become a topic for a separate follow-up work. Researchers interested in hybrid models may consider comparing a small set of learners and optimizers under consistent validation protocols and may also investigate sensitivity to learning rates and hyperparameter ranges.
The use of hybrid methodologies has become widespread across many disciplines today. Recently, deep learning techniques such as CNN, and CNN-LSTM have gained increasing visibility in research. These methods are considered highly innovative due to their nature as modifications of ANN architecture119. However, in this study, a key motivation was to benchmark the extent to which an established method (SVM) can benefit from input-structure design and metaheuristic tuning in a high-latitude PV forecasting context before moving to more complex deep-learning approaches. Deep-learning methods remain a promising direction for future work.
The models utilized in this study are data-driven models. Although data-driven models have advanced beyond physical-based models, the performance comparison between the two modeling approaches is frequently debated among researchers. This debate arises from the fact that physical-based models incorporate multiple parameters that accurately portray real-life phenomena. In contrast data-driven models can be very powerful but they also rely on heavily the quality and quantity of available data and sometimes may not capture certain physical constraints. Therefore, for critical applications, both approaches should be used to cross-validate results and ensure robustness for the area being examined.
Limitations of this study include: (i) daily aggregation of PV output, which smooths intra-day dynamics and ramp events; (ii) the use of a fixed 7-hour daytime window, which may not fully represent seasonal daylight variability at high latitudes; (iii) reliance on PVGIS-ERA5 derived PV output rather than site-measured power, which can introduce modeling biases; (iv) a restricted predictor set (primarily temperature and lagged PV) without additional meteorological drivers (e.g., irradiance, cloud cover, wind); (v) a limited number of sites and years (2020–2023); and (vi) evaluation based on a single chronological split without rolling-origin cross-validation or probabilistic forecasts.
These limitations restrict generalization beyond the studied sites and period. Future work should (i) use measured PV data when available, (ii) expand predictors, (iii) benchmark against additional ML/DL baselines (e.g., RF, gradient boosting, LSTM/GRU, CNN-LSTM), (iv) apply rolling-origin evaluation, and (v) report prediction intervals to quantify forecast uncertainty. Addressing these limitations would allow to generalize the findings.
This study developed SVM-based hybrid models tuned with ABC, GWO, GA, and PSO to forecast daily PV power for three southern Norway locations (Kristiansand, Stavanger, and Oslo) using PVGIS-ERA5 derived PV output and temperature data covering 2020–2023. A chronological 70/30 train–test split was used to evaluate out-of-sample performance.
Across all sites and input configurations, metaheuristic tuning improved test performance relative to the untuned SVM baseline. PSO produced the best mean test metrics, although differences among the top-performing optimizers were generally small.
Input configurations that included both lagged PV power and lagged temperature (M03–M04) typically outperformed PV-only structures (M01–M02), confirming the value of incorporating meteorological inputs in PV forecasting.
Model skill varied by location, with the highest scores obtained for Kristiansand in this dataset. These results should be interpreted as case-study findings for southern Norway rather than universal performance rankings.
Visual error analysis (e.g., Bland–Altman plots) plots were particularly helpful for diagnosing bias and agreement between observations and predictions; however, when model metrics are very close, visual methods alone may be insufficient to clearly distinguish models.
However, this study’s scope is limited to daily data from three sites and a specific train-test split; thus, the results should be interpreted as a case study rather than generalized truths.
The use of simulated PVGIS data, a restricted set of input features, and absence of cross-validation can be list as key limitations in this study and will be addressed in future work.
Overall, the proposed SVM–metaheuristic framework provides a benchmark for high-latitude PV power forecasting. It demonstrates that even established machine learning models can benefit significantly from intelligent input selection and parameter tuning. Future studies can build on this by incorporating additional data sources (e.g., irradiation and cloud cover measurements) and exploring advanced models to further improve forecast reliability.
The hyper-parameters used for each technique are shown in Table 5.
The original contributions presented in the study are included in the article. The raw data supporting the conclusions of this article will be made available by the corresponding author upon reasonable request.
Bosetti, V., Catenacci, M., Fiorese, G. & Verdolini, E. The future prospect of PV and CSP solar technologies: an expert elicitation survey. Energy Policy. 49, 308–317 (2012).
Gernaat, D. E. et al. Climate change impacts on renewable energy supply. Nat. Clim. Change. 11, 119–125 (2021). van.
Sharma, V. K. et al. Imperative role of photovoltaic and concentrating solar power technologies towards renewable energy generation. Int. J. Photoenergy 2022, 3852484 (2022).
Wild, M., Folini, D., Henschel, F., Fischer, N. & Müller, B. Projections of long-term changes in solar radiation based on CMIP5 climate models and their influence on energy yields of photovoltaic systems. Sol. Energy. 116, 12–24 (2015).
Jerez, S. et al. The impact of climate change on photovoltaic power generation in Europe. Nat. Commun. 6, 10014 (2015).
Article  ADS  CAS  PubMed  Google Scholar 
Zou, L. et al. Global surface solar radiation and photovoltaic power from coupled model intercomparison project phase 5 climate models. J. Clean. Prod. 224, 304–324 (2019).
Müller, J., Folini, D., Wild, M. & Pfenninger, S. CMIP-5 models project photovoltaics are a no-regrets investment in Europe irrespective of climate change. Energy 171, 135–148 (2019).
Bichet, A. et al. Potential impact of climate change on solar resource in Africa for photovoltaic energy: analyses from CORDEX-AFRICA climate experiments. Environ. Res. Lett. 14, 124039 (2019).
Park, C. et al. What determines future changes in photovoltaic potential over East asia? Renew. Energy. 185, 338–347 (2022).
Dutta, R., Chanda, K. & Maity, R. Future of solar energy potential in a changing climate across the world: A CMIP6 multi-model ensemble analysis. Renew. Energy. 188, 819–829 (2022).
Dubey, S., Sarvaiya, J. N. & Seshadri, B. Temperature dependent photovoltaic (PV) efficiency and its effect on PV production in the world–a review. Energy Procedia. 33, 311–321 (2013).
Sawadogo, W., Abiodun, B. J. & Okogbue, E. C. Impacts of global warming on photovoltaic power generation over West Africa. Renew. Energy. 151, 263–277 (2020).
Duran, Y. et al. Investigation of the near future solar energy changes using a regional climate model over Istanbul. Türkiye Energies. 17, 2644 (2024).
Udo, W. S., Kwakye, J. M., Ekechukwu, D. E. & Ogundipe, O. B. Predictive analytics for enhancing solar energy forecasting and grid integration. Eng. Sci. Technol. J. 4, 589–602 (2023).
Kaur, A. Forecasting for Power Grids with High Solar Penetration (University of California, 2015).
Dzobo, O. & Kuntwane, T. Prediction of Solar PV power output using neural networks. in IEEE International Conference on Internet of Things and Intelligence Systems (IoTaIS) (IEEE, 2024).
Deshmukh, S., Limkar, S., Nagthane, R., Pande, V. & Tare, A. V. Design of grid-connected solar PV system integrated with battery energy storage system. in 3rd Asian Conference on Innovation in Technology (ASIANCON) (IEEE, 2023).
Clifort, R. & Ali, A. Investigation of machine learning for predicting the output of photovoltaic solar power. in Global Energy Conference (GEC) (IEEE, 2024).
Bassey, K. E. Solar energy forecasting with deep learning technique. Eng. Sci. Technol. J. 4, 18–32 (2023).
Article  Google Scholar 
Aupke, P., Kassler, A., Theocharis, A., Nilsson, M. & Uelschen, M. Quantifying uncertainty for predicting renewable energy time series data using machine learning. Eng. Proc. 5, 50 (2021).
Paulescu, M. & Paulescu, E. Short-term forecasting of solar irradiance. Renew. Energy. 143, 985–994 (2019).
Zambrano, A. F. & Giraldo, L. F. Solar irradiance forecasting models without on-site training measurements. Renew. Energy. 152, 557–566 (2020).
Akarslan, E., Hocaoglu, F. O. & Edizkan, R. Novel short term solar irradiance forecasting models. Renew. Energy. 123, 58–66 (2018).
Diagne, M., David, M., Lauret, P. Boland, J. & Schmutz, N. Review of solar irradiance forecasting methods and a proposition for small-scale insular grids. Renew. Sustain. Energy Rev. 27, 65–76 (2013).
Ajith, M. & Martínez-Ramón, M. Deep learning algorithms for very short term solar irradiance forecasting: A survey. Renew. Sustain. Energy Rev. 182, 113362 (2023).
Wang, F. et al. A minutely solar irradiance forecasting method based on real-time Sky image-irradiance mapping model. Energy. Conv. Manag. 220, 113075 (2020).
Caldas, M. & Alonso-Suárez, R. Very short-term solar irradiance forecast using all-sky imaging and real-time irradiance measurements. Renew. Energy. 143, 1643–1658 (2019).
Miller, S. D., Rogers, M. A., Haynes, J. M., Sengupta, M. & Heidinger, A. K. Short-term solar irradiance forecasting via satellite/model coupling. Sol. Energy. 168, 102–117 (2018).
Wang, F., Mi, Z. & Su, S. Zhao Short-Term solar irradiance forecasting model based on artificial neural network using statistical feature parameters. Energies 5, 1355–1370 (2012).
Article  Google Scholar 
Ağbulut, Ü., Gürel, A. E. & Biçen, Y. Prediction of daily global solar radiation using different machine learning algorithms: evaluation and comparison. Renew. Sustain. Energy Rev. 135, 110114 (2021).
Huang, X. et al. Hybrid deep neural model for hourly solar irradiance forecasting. Renew. Energy. 171, 1041–1060 (2021).
Dou, W. et al. Sreeram A multi-modal deep clustering method for day-ahead solar irradiance forecasting using ground-based cloud imagery and time series data. Energy 321, 135285 (2025).
Article  Google Scholar 
Alzahrani, A., Shamsi, P., Dagli, C. & Ferdowsi, M. Solar irradiance forecasting using deep neural networks. Procedia Comput. Sci. 114, 304–313 (2017).
Yu, Y., Cao, J. & Zhu, J. An LSTM Short-Term solar irradiance forecasting under complicated weather conditions. IEEE Access. 7, 1–1 (2019).
Brahma, B. & Wadhvani, R. Solar irradiance forecasting based on deep learning methodologies and Multi-Site data. Symmetry 12, 1830 (2020).
Alorf, A. & Khan, M. U. G. Solar irradiance forecasting using temporal fusion transformers. Int. J. Energy Res. 2025, 3534500 (2025).
Kumari, P. & Toshniwal, D. Deep learning models for solar irradiance forecasting: A comprehensive review. J. Clean. Prod. 318, 128566 (2021).
Oruc, S., Tugrul, T. & Hinis, M. A. Beyond traditional metrics: exploring the potential of hybrid algorithms for drought characterization and prediction in the Tromso Region, Norway. Appl. Sci. 14, 7813 (2024).
Zhang, Y., Wang, J., Li, X. & Huang, S. Wang feature selection for high-dimensional datasets through a novel artificial bee colony framework. Algorithms 14, 324 (2021).
Article  Google Scholar 
Gujarathi, P. K., Shah, V. A. & Lokhande, M. M. Hybrid artificial bee colony-grey Wolf algorithm for multi-objective engine optimization of converted plug-in hybrid electric vehicle. Adv. Energy Res. 35–52 (2020).
Mustaffa, Z., Sulaiman, M. H. & Kahar, M. N. M. LS-SVM hyper-parameters optimization based on GWO algorithm for time series forecasting. in 4th international conference on software engineering and computer systems (ICSECS) (IEEE, 2015).
Ahmad, I. F., Qayum, S. U., Rahman, G. & Srivastava, G. Using improved hybrid grey Wolf algorithm based on artificial bee colony algorithm onlooker and scout bee operators for solving optimization problems. Int. J. Comput. Intell. Syst. 17, 111 (2024).
Liang, J. et al. Using adaptive chaotic grey Wolf optimization for the daily streamflow prediction. Expert Syst. Appl. 237, 121113 (2024).
Holland, J. H. Adaptation in Natural and Artificial Systems: an Introductory Analysis with Applications To biology, control, and Artificial Intelligence (MIT Press, 1992).
Xian, Z., Xie, J. & Wang, Y. Representative artificial bee colony algorithms: a survey. in LISS 2012: Proceedings of 2nd International Conference on Logistics, Informatics and Service Science. Springer. (2013).
Elyan, E. & Gaber, M. M. A genetic algorithm approach to optimising random forests applied to class engineered data. Inf. Sci. 384, 220–234 (2017).
Sircar, A., Yadav, K., Rayavarapu, K., Bist, N. & Oza, H. Application of machine learning and artificial intelligence in oil and gas industry. Petroleum Res. 6, 379–391 (2021).
Mirjalili, S. & Lewis, A. The Whale optimization algorithm. Adv. Eng. Softw. 95, 51–67 (2016).
Tang, C. et al. A hybrid Whale optimization algorithm with artificial bee colony. Soft. Comput. 26, 2075–2097 (2022).
Pierezan, J. & Coelho, L. D. S. Coyote optimization algorithm: a new metaheuristic for global optimization problems. in 2018 IEEE congress on evolutionary computation (CEC). IEEE. (2018).
Chinh, N. C., Tung, N. N. & Thuan, N. Q. Coyote Optimization Algorithm-Based PV Planning Strategy for Maximizing Hosting Capacity in A Distribution System. in 2023 Asia Meeting on Environment and Electrical Engineering (EEE-AM) ( IEEE, 2023).
Chin, V. J. & Salam, Z. Coyote optimization algorithm for the parameter extraction of photovoltaic cells. Sol. Energy. 194, 656–670 (2019).
Tong, H., Zhu, Y., Pierezan, J., Xu, Y. & Coelho, L. d. S. Chaotic Coyote optimization algorithm. J. Ambient Intell. Humaniz. Comput. 13, 2807–2827 (2022).
Pierezan, J., Coelho, L., Mariani, V. & Lebensztajn, L. Multiobjective coyote algorithm applied to electromagnetic optimization. In: 2019 22nd international conference on the computation of electromagnetic fields (COMPUMAG). IEEE, 1–4 (2019).
Myhre, S. F. & Rosenberg, E. The role and impact of rooftop photovoltaics in the norwegian energy system under different energy transition pathways. Advanced Energy and Sustainability Research 2400184 (2025).
Gholami, H. Technical potential of solar energy in buildings across norway: capacity and demand. Sol. Energy. 278, 112758 (2024).
Article  Google Scholar 
Meng, Q., Yin, H., Johannessen, M. R. & Berntzen, L. Analysis of development of Norwegian household solar energy ecosystem. in Journal of Physics: Conference Series. IOP Publishing. (2023).
Abdeldayem, M. M., Aldulaimi, S. H., Al-Kaabi, H., Baqi, A. & Muttar, A. K. Empowering Greener Horizons: Novel Stakeholder Engagement Impact on Norway’s Solar-PV Storage Adoption. in 2024 ASU International Conference in Emerging Technologies for Sustainability and Intelligent Systems (ICETSIS). IEEE. (2024).
Skogsberg, L. What are the past and future trends of solar energy in Norway? A study on Norway’s solar energy system. (Dissertation) KTH, School of Industrial Engineering and Management (2024).
Yordanov, G. H. Characterization and analysis of photovoltaic modules and the solar resource based on in-situ measurements in southern Norway. Norwegian University of Science and Technology, Department of Electric Power Engineering, Trondheim (2012).
Øgaard, M. B., Riise, H. N., Haug, H., Sartori, S. & Selj, J. H. Photovoltaic system monitoring for high latitude locations. Sol. Energy. 207, 1045–1054 (2020).
Skoplaki, E. & Palyvos, J. A. On the temperature dependence of photovoltaic module electrical performance: A review of efficiency/power correlations. Sol. Energy. 83, 614–624 (2009).
Hudisteanu, V. S., Chereches, N. C., Turcanu, F. E., Hudisteanu, I. & Romila, C. Impact of temperature on the efficiency of monocrystalline and polycrystalline photovoltaic panels: A comprehensive experimental analysis for sustainable energy solutions. Sustainability 16, 10566 (2024).
Manni, M., Failla, M. C., Nocente, A., Lobaccaro, G. & Jelle, B. P. The influence of Icephobic nanomaterial coatings on solar cell panels at high latitudes. Sol. Energy. 248, 76–87 (2022).
Ghazi, A., Barahmand, Z. & Øi, L. E. The effect of climate and orientation on the energy performance of a prefab house in Norway. Norway. Scandinavian Simulation Society, 2023, 62–69., 62–69 (2023).
Rees, G., Hebryn-Baidy, L. & Good, C. Estimating the potential for rooftop generation of solar energy in an urban context using High-Resolution open access Geospatial data: A case study of the City of Tromsø, Norway. ISPRS Int. J. Geo-Information. 14, 123 (2025).
Khanna, S., Reddy, K. & Mallick, T. K. Performance analysis of Tilted photovoltaic system integrated with phase change material under varying operating conditions. Energy 133, 887–899 (2017).
Kashan, M. E. & Fung, A. & J. Swift Integrating Novel Microchannel-Based Solar Collectors with a Water-to-Water Heat Pump for Cold-Climate Domestic Hot Water Supply, Including Related Solar Systems Comparisons. Energies. 14, 4057 (2021).
Barbosa, J. et al. Tilt correction to maximize energy yield from bifacial PV modules. in IOP Conference Series: Earth and Environmental Science (IOP Publishing, 2022).
Ogundimu, E. O., Akinlabi, E., Mgbemene, C. & Jacobs, I. Designing and fabrication of an installation PV solar modules tilting platform. Sustain. Eng. Innov. 4, 46–57 (2022).
Tuğrul, T. & Hınıs, M. A. Oruç comparison of LSTM and SVM methods through wavelet decomposition in drought forecasting. Earth Sci. Inf. 18, 139 (2025).
Article  ADS  Google Scholar 
Tuğrul, T. & Hinis, M. A. Improvement of drought forecasting by means of various machine learning algorithms and wavelet transformation. Acta Geophys. 73, 855–874 (2025).
Tuğrul, T. et al. Data-Driven drought prediction by means of machine learning techniques and increasing accuracy with wavelet transform. Pure. appl. Geophys. 182, 4319–4341 (2025).
Rezaiy, R. & Shabri, A. Integrating wavelet transform and support vector machine for improved drought forecasting based on standardized precipitation index. J. Hydroinformatics. 27, 320–337 (2025).
Taylor, N. et al. Photovoltaics Geographical Information System: Status Report 2024 (2025).
Ding, C. & Peng, H. Minimum redundancy feature selection from microarray gene expression data. J. Bioinform. Comput. Biol. 3, 185–205 (2005).
Latifoğlu, L. & Kaya, E. High-performance prediction model combining minimum redundancy maximum relevance, circulant spectrum analysis, and machine learning methods for daily and peak streamflow. Theoret. Appl. Climatol. 155, 621–643 (2024).
Ma, X., Zhang, Y. & Wang, Y. Performance evaluation of kernel functions based on grid search for support vector regression. in IEEE 7th international conference on cybernetics and intelligent systems (CIS) and IEEE conference on robotics, automation and mechatronics (RAM) (2015).
Zhang, Z. Customizing Kernels in Support Vector Machines (Master’s thesis) University of Waterloo (2007).
Fathi Hafshejani, S. & Moaberfard, Z. A new trigonometric kernel function for support vector machine. Iran. J. Comput. Sci. 6, 137–145 (2023).
Lu, D. G. & G.-B. Li Reproducing kernel-based support vector machine for structural reliability analysis 12th International Conference on Applications of Statistics and Probability in Civil Engineering (ICASP) (2015).
Wang, W., Xu, Z., Lu, W. & Zhang, X. Determination of the spread parameter in the Gaussian kernel for classification and regression. Neurocomputing 55, 643–663 (2003).
Kusuma, M. P. & Kudus, A. Penerapan Metode Support Vector Regression (SVR) pada Data Survival KPR PT. Bank ABC, Tbk. in Bandung Conference Series: Statistics. (2022).
Wang, S., Deng, Z., Chung, F. -I. & Hu, W. From Gaussian kernel density Estimation to kernel methods. Int. J. Mach. Learn. Cybernet. 4, 119–137 (2013).
Oruc, S., Hınıs, M. A., Selek, Z. & Tuğrul, T. Deep signals: enhancing bottom temperature predictions in norway’s Mjøsa lake through VMD- and EMD-Boosted machine. Learn. Models Water. 17, 2673 (2025).
Karaboga, D. An idea based on honey bee swarm for numerical optimization Technical report-tr06, Erciyes University (2005).
Siva Kumar, M. et al. A hybrid approach of ANFIS—artificial bee colony algorithm for intelligent modeling and optimization of plasma Arc cutting on Monel™ 400 alloy. Materials 14, 6373 (2021).
Li, Z., Wang, W., Yan, Y. & Li, Z. PS–ABC: A hybrid algorithm based on particle swarm and artificial bee colony for high-dimensional optimization problems. Expert Syst. Appl. 42, 8881–8895 (2015).
Article  Google Scholar 
Vitorino, L. Ribeiro, S. F. & Bastos-Filho, C. J. A mechanism based on artificial bee colony to generate diversity in particle swarm optimization. Neurocomputing 148, 39–45 (2015).
Xu, F. et al. A new global best guided artificial bee colony algorithm with application in robot path planning. Appl. Soft Comput. 88, 106037 (2020).
Mirjalili, S., Mirjalili, S. M. & Lewis, A. Grey Wolf Optimizer. Advances in Engineering Software 69, 46–61 (2014).
Jaafari, A. et al. Meta optimization of an adaptive neuro-fuzzy inference system with grey Wolf optimizer and biogeography-based optimization algorithms for Spatial prediction of landslide susceptibility. Catena 175, 430–445 (2019).
Muro, C., Escobedo, R., Spector, L. & Coppinger, R. Wolf-pack (Canis lupus) hunting strategies emerge from simple rules in computational simulations. Behav. Process. 88, 192–197 (2011).
Hao, P. & Sobhani, B. Application of the improved chaotic grey Wolf optimization algorithm as a novel and efficient method for parameter Estimation of solid oxide fuel cells model. Int. J. Hydrog. Energy 46, 36454–36465 (2021).
Dehghani, M., Seifi, A. & Riahi-Madvar, H. Novel forecasting models for immediate-short-term to long-term influent flow prediction by combining ANFIS and grey Wolf optimization. J. Hydrol. 576, 698–725 (2019).
Mirboluki, A., Mehraein, M. & Kisi, O. Improving accuracy of neuro fuzzy and support vector regression for drought modelling using grey Wolf optimization. Hydrol. Sci. J. 67, 1582–1597 (2022).
Oladipo, S. & Sun, Y. Enhanced adaptive neuro-fuzzy inference system using genetic algorithm: A case study in predicting electricity consumption. SN Appl. Sci. 5, 186 (2023).
Hong, H. et al. Applying genetic algorithms to set the optimal combination of forest fire related variables and model forest fire susceptibility based on data mining models. The case of Dayu County, China. Sci. Total Environ. 630, 1044–1056 (2018).
Zhang, G., Wu, M., Duan, W. & Huang, X. Genetic algorithm based QoS perception routing protocol for VANETs. Wirel. Commun. Mob. Comput. 2018, 3897857 (2018).
Brás, G., Silva, A. M. & Wanner, E. F. A genetic algorithm for rule extraction in fuzzy adaptive learning control networks. Genet. Program Evolvable Mach. 25, 11 (2024).
Zanganeh, M. Simultaneous optimization of clustering and fuzzy IF-THEN rules parameters by the genetic algorithm in fuzzy inference system-based wave predictor models. J. Hydroinform. 19, 385–404 (2017).
Article  Google Scholar 
Khanmohammadi, S., Kizilkan, O. & Musharavati, F. Multiobjective Optimization of a Geothermal Power plant, in Thermodynamic Analysis and Optimization of Geothermal Power Plants 279–291 (Elsevier, 2021).
Martinez, C. M. & Cao, D. Integrated energy management for electrified vehicles iHorizon-Enabled Energy Management for Electrified Vehicles (Butterworth-Heinemann, (2019).
Saha, S. et al. Proposing novel ensemble approach of particle swarm optimized and machine learning algorithms for drought vulnerability mapping in Jharkhand, India. Geocarto Int. 37, 8004–8035 (2022).
Achite, M. et al. Hydrological drought prediction based on hybrid extreme learning machine: Wadi Mina basin case study. Algeria Atmosphere. 14, 1447 (2023).
Cheng, C., Sha, Q., He, B. & Li, G. Path planning and obstacle avoidance for AUV: A review. Ocean Eng. 235, 109355 (2021).
Kennedy, J. & Eberhart, R. Particle swarm optimization. in Proceedings of ICNN’95-international conference on neural networks (1995).
Wang, D., Tan, D. & Liu, L. Particle swarm optimization algorithm: an overview. Soft. Comput. 22, 387–408 (2018).
Article  Google Scholar 
Rajabi Kuyakhi, H. & Tahmasebi-Boldaji, R. Developing an adaptive neuro-fuzzy inference system based on particle swarm optimization model for forecasting Cr(VI) removal by NiO nanoparticles. Environ Prog Sustainable Energy 40 (2021).
Eshaghzadeh, A. & Hajian, A. 2-D gravity inverse modelling of anticlinal structure using improved particle swarm optimization (IPSO). Arab. J. Geosci. 14, 1378 (2021).
Article  Google Scholar 
Nimmanterdwong, P., Chalermsinsuwan, B. & Piumsomboon, P. Optimizing utilization pathways for biomass to chemicals and energy by integrating emergy analysis and particle swarm optimization (PSO). Renew. Energy 202, 1448–1459 (2023).
Article  Google Scholar 
Rahmati, O. et al. Capability and robustness of novel hybridized models used for drought hazard modeling in Southeast Queensland, Australia. Sci. Total Environ. 718, 134656 (2020).
Article  CAS  PubMed  Google Scholar 
Shorabeh, S. N. et al. A decision model based on decision tree and particle swarm optimization algorithms to identify optimal locations for solar power plants construction in Iran. Renew. Energy. 187, 56–67 (2022).
Article  Google Scholar 
Kashyap, N. & Mishra, A. A discourse on metaheuristics techniques for solving clustering and semisupervised learning models. in Cognitive Big Data Intelligence with a Metaheuristic Approach 1–19 (Elsevier, 2022).
Knoben, W. J., Freer, J. E. & Woods, R. A. Inherent benchmark or not? Comparing Nash–Sutcliffe and Kling–Gupta efficiency scores. Hydrol. Earth Syst. Sci. 23, 4323–4331 (2019).
Article  ADS  Google Scholar 
AlMohimeed, A. et al. ViT-PSO-SVM: Cervical cancer predication based on integrating vision transformer with particle swarm optimization and support vector machine. Bioengineering 11, 729 (2024).
Du, G. & Ou, R. Integrated support vector machine with improved PSO optimization for early risk screening and prevention of stroke in patients with hypertension. Comput. Ind. Eng. 207, 111300 (2025).
Article  Google Scholar 
Samantaray, S., Sahoo, A. & Agnihotri, A. Prediction of flood discharge using hybrid PSO-SVM algorithm in Barak river basin. MethodsX 10, 102060 (2023).
Article  PubMed  PubMed Central  Google Scholar 
Kaleybar, F. A. & Molavi, A. CNN-LSTM-RF integration for predicting Mississippi river discharge dynamics. Acta Geophys. 73, 6005–6020 (2025).
Article  ADS  Google Scholar 
Download references
Funding: This research received no external funding. APC was supported/funded by UiT, the Arctic University of Norway. Acknowledgments: During the preparation of this work, the authors used ChatGPT and Quillbot in order to improve readability, edit grammar, and language of some parts of the manuscript. After using this tool/service, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication. Conflicts of Interest: The authors declare that there is no conflict of interest.
Open access funding provided by UiT The Arctic University of Norway (incl University Hospital of North Norway). This research received no external funding. APC was supported/funded by UiT, the Arctic University of Norway.
The Center for Sámi Studies, UiT Norges Arktiske Universitet, Tromsø, 9037, Norway
Sertaç Oruç
Civil Engineering Department, Faculty of Engineering and Natural Sciences, Ankara Yıldırım Beyazıt University, 15 Temmuz Şehitleri Campus, Ankara, 06010, Turkey
Sertaç Oruç
Civil Engineering Department, Faculty of Engineering, Central Campus, Aksaray University, Aksaray, 68100, Türkiye, Turkey
Mehmet Ali Hınıs
Civil Engineering Department, Technology Faculty, Central Campus, Gazi University, Ankara, 06560, Turkey
Türker Tuğrul
PubMed Google Scholar
PubMed Google Scholar
PubMed Google Scholar
Conceptualization, S.O. and M.A.H.; Formal analysis, T.T.; Methodology, T.T., S.O. and M.A.H.; Supervision, S.O. and M.A.H.; Visualization, T.T., S.O. and M.A.H.; Writing—original draft, T.T. and S.O.; Writing—review and editing, T.T., S.O., and M.A.H. All authors have read and agreed to the published version of the manuscript.
Correspondence to Sertaç Oruç.
The authors declare no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Reprints and permissions
Oruç, S., Hınıs, M.A. & Tuğrul, T. Forecasting photovoltaic power in high-latitude regions via support vector machine optimized by meta-heuristics. Sci Rep 16, 3438 (2026). https://doi.org/10.1038/s41598-025-33415-7
Download citation
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-025-33415-7
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative
Advertisement
Scientific Reports (Sci Rep)
ISSN 2045-2322 (online)
© 2026 Springer Nature Limited
Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.

source

This entry was posted in Renewables. Bookmark the permalink.

Leave a Reply